SlideShare uma empresa Scribd logo
1 de 29
Knowledge-Primed Neural Networks Enable
Biologically Interpretable Deep Learning on
Single-Cell Sequencing Data
Course: Machine Learning and Drug Design
Nikolaus Fortelny and Chistoph Bock
Knowledge-Primed Neural Networks
(KPNNs)
6/3/2023 5:46 PM 2
Trained ANNs typically lack interpretability, i.e., the ability to provide
human-understandable, high-level explanations of how they transform inputs
(prediction attributes) into outputs which is a major limitation to the wider
application of deep learning in biology.
Figure 1: ANN vs. knowledge-primed neural networks (KPNNs)
Introduction Results & Discussion
Methodology Conclusion
Methodology
6/3/2023 5:46 PM 3
Figure 2: Pipeline for the construction of KPNN
Construction of general regulatory neural networks for human cell
Construction of application specific KNNs for deep learning
Applying Deep Learning Approach to KNNs
Optimization of Learning Method to Enhance KPNN interpretability
Training of the optimized KPNN using single cell RNA seq data
Comparing ANN and KPNN
Introduction Results & Discussion
Methodology Conclusion
6/3/2023 5:46 PM 4
1.Constructing Generic KPNNs
• Integrates multiple signaling pathways and gene-regulatory interactions into
a single network.
• For the biological basis, large-scale, genome-wide datasets of transcription
factor target genes and protein signaling interactions from several public
databases were integrated into a broad network model of the potential
regulatory space in human cells.
• This general regulatory network model is represented as a directed graph.
Introduction Results & Discussion
Methodology Conclusion
6/3/2023 5:46 PM 5
1.Constructing Generic KPNNs
Introduction Results & Discussion
Methodology Conclusion
Figure 3: Generic acyclic graphs
6/3/2023 5:46 PM 6
2. Construction of Application-Specific
KPNNs for Deep Learning
• The resulting graphs are used for interpretable deep learning:
Measured gene expression values (input nodes) predict the regulatory
importance of connected transcription factors (hidden nodes), which predict
the regulatory importance of their connected signaling proteins (hidden nodes),
which then predict the regulatory importance and activation states of surface
receptors, which finally predict the output (phenotype) of the specific dataset.
 Two KPNNs were constructed in this study:
1. TCR signaling (TCR KPNN), to predict TCR stimulation from gene
expression
2. Generalized KPNN (GEN KPNN) to predict cell states from single-cell
RNA-seq datasets independent of specific receptors.
Introduction Results & Discussion
Methodology Conclusion
6/3/2023 5:46 PM 7
2. Construction of Application-Specific
KPNNs for Deep Learning
G1
G2
G3
TF1
TF2
TF3
TF4
SP1
SP 2
SP 3
SP 4
0
1
Objective Database Used
TF- Target Gene
Pair
Harmonizome & TTRUST
Signaling Proteins
– Target Proteins
SIGNOR
Introduction Results & Discussion
Methodology Conclusion
Figure 4: Application specific KPNNs
Table 1:Databases used for dataset acquisition
6/3/2023 5:46 PM 8
3. Datasets Used for Training KPNN
Interpretable deep learning on KPNNs was developed and evaluated using
simulated transcriptome profiles.
 It was validated and benchmarked on the biological data of TCR
stimulation.
Model was further validated on 4 benchmark datasets
1. HCA
2. LCH
3. AML
4. Glioblastoma
Introduction Results & Discussion
Methodology Conclusion
3. Datasets Used for Training KPNN
6/3/2023 5:46 PM 9
• TCR Dataset
downloaded from GEO (GSE137554).
Class Values: TCR stimulated (1) – TCR un-stimulated samples (0)
• Simulated Dataset
We averaged expression counts for the unstimulated state for each gene to
generate a baseline expression profile.
 a positive or negative twofold change was introduced in selected genes
resulting in a second (differential) gene expression profile.
Class Values: TCR stimulated (1) – TCR un-stimulated samples (0).
Introduction Results & Discussion
Methodology Conclusion
4. Applying Deep learning to KPNNs
6/3/2023 5:46 PM 10
1.
Preprocessing
the input data
2.
Training of the
KPNN or ANN
deep neural
networks
3.
Prediction
evaluation
Introduction Results & Discussion
Methodology Conclusion
Figure 5: Methodology of generic deep learning applied to KPNNs
4.1. Preprocessing of Input
6/3/2023 5:46 PM 11
• Input data were split into training set (60%), validation set (20%), and
test set (20%) using numpy random.choice.
• Gene expression data were converted to log (TPM + 1) values and
normalized at the scale of 0 -1.
• Normalization factors were calculated based on the training data and
then applied to the validation and test data.
Introduction Results & Discussion
Methodology Conclusion
4.2. Network Training
6/3/2023 5:46 PM 12
Parameters Configurations
Implementation Tensor flow
Edge weights Randomly initialized
Activation function Sigmoid
Loss Function A weighed cross entropy with L2 regularization.
Weighted class entropy- to improve training in the presence of class imbalance, the cross-
entropy of each sample was weighted by the number of samples of
each class.
Class Weight (1/Nclasses)/(Nx/Nsamples)
where Nclasses is the number of classes, Nx is the number of samples of
class x, and Nsamples is the total number of samples
Introduction Results & Discussion
Methodology Conclusion
Table 2: Configurations for training KPNNs
4.2. Network Training
6/3/2023 5:46 PM 13
• Error rate: True label probability – Predicted class probability
• Training was terminated with early stopping, which was triggered by a
specific number (“patience”) of epochs without considerable learning
processes (“failed epochs”).
• Epochs were considered as “failed epochs” if either
(i) loss does not show significant improvement and remains relatively constant over a
certain number of epochs.
(ii) validation set error increased (indicative of overfitting)
• Models were saved during training if they achieved considerable learning
progress compared to the previously saved model.
Introduction Results & Discussion
Methodology Conclusion
5. Optimization of Learning Method to
Enhance KPNN interpretability
6/3/2023 5:46 PM 14
• Three modifications were implemented to enhance biological
interpretability of the trained KPNNs:
dropout on hidden nodes to improve robustness.
dropout on input nodes to increase quantitative interpretability.
training on simulated control inputs to normalize for uneven
connectivity in biological networks.
Introduction Results & Discussion
Methodology Conclusion
6.Training the KPNN on the Single
Cell RNA Seq. Dataset
6/3/2023 5:46 PM 15
• Raw read counts of all 1735 cells, class values (stimulated or
unstimulated), and an edge list encoding the TCR KPNN structure
were provided as input.
• Learning was stopped after 20 failed epochs.
• During learning, new models were saved if they reduced the validation
set error of the latest stored (i.e., so far best-performing) model by at
least 20%.
• Dropout rates from 0 to 50% were evaluated in steps of 10 percentage
points.
• Network training was repeated 90 times for each level of dropout
Introduction Results & Discussion
Methodology Conclusion
Results and Discussion
Maria Hassan Khan
6/3/2023 5:46 PM 16
Introduction Results & Discussion
Methodology Conclusion
KPNNs Establish Deep Learning on
Biological Networks
6/3/2023 5:46 PM 17
• Constructed KPNNs reflected the typical flow of information in cells:
Signals are transduced from receptors via signaling proteins to transcription
factors, which in turn induce changes in gene expression.
Introduction Results & Discussion
Methodology Conclusion
Figure 6: Mapping genome-wide regulatory networks to KPNNs
KPNNs Establish Deep Learning on
Biological Networks
6/3/2023 5:46 PM 18
• KPNNs can be trained based on single-cell RNA-seq data for cells in different
states (e.g., subtypes of cancer and immune cells) or under different environmental
exposures (e.g., receptor-stimulated vs unstimulated cells)
• After successful completion of the learning process, each fitted edge weight is
taken as an indicator for the relevance of the corresponding regulatory relationship
in the investigated biological system.
• We then derive node weights from the edge weights, in order to identify signaling
proteins and transcription factors that are likely relevant in this biological system.
• KPNNs thus exploit the ability of deep learning algorithms to assign meaningful
weights across multiple hidden layers in KPNNs, thereby identifying and
prioritizing relevant regulatory proteins for experimental validation and biological
interpretation.
Introduction Results & Discussion
Methodology Conclusion
TCR Signaling
• KPNNs predicted TCR stimulation
with a median receiver operating
characteristic (ROC) area under
curve (AUC) value of 0.984
whereas, ANN achieved value
0.948.
6/3/2023 5:46 PM 19
Introduction Results & Discussion
Methodology Conclusion
Figure 7: Schematic diagram of TCR KPNN
6/3/2023 5:46 PM 20
KPNN Leading to Interpretable Deep
Learning
Introduction Results & Discussion
Methodology Conclusion
Figure 8: Comparison of the number of layers in TCR KPNN vs. TCR ANN
6/3/2023 5:46 PM 21
KPNN Leading to Interpretable Deep
Learning
Introduction Results & Discussion
Methodology Conclusion
Figure 9: Comparison of the number of degrees of different layers in TCR KPNN vs.
TCR ANN
6/3/2023 5:46 PM 22
Interpretability of KPNN Network
Connectedness
Introduction Results & Discussion
Methodology Conclusion
Figure 10: Comparison of the number of outdegrees of different layers in TCR KPNN vs.
6/3/2023 5:46 PM 23
Interpretability of KPNN Network
Connectedness
Introduction Results & Discussion
Methodology Conclusion
Figure 11: Comparison of the network connectedness of TCR KPNN vs. TCR ANN
6/3/2023 5:46 PM 24
Enhanced KPNN Interpretability via
Optimized Learning Methods
• Application of generic deep learning to KPNNs instead of ANNs,
yielded high prediction accuracies with three recurrent challenges in
interpreting the learned models:
low reproducibility in the presence of network redundancy
lack of quantitative interpretability (i.e., node weights fail to quantitatively
reflect the predictive power of individual nodes)
uneven connectivity inherent to biological networks, which affects node
weights independent of training data
Introduction Results & Discussion
Methodology Conclusion
6/3/2023 5:46 PM 25
Enhanced KPNN Interpretability via
Optimized Learning Methods
1. Hidden node dropout.
2. Input dropout.
3. Correction for network structure.
• These three methodological optimizations enhanced the biological
interpretability of KPNNs both on simulated and on real data.
• It is anticipated that similar optimizations will also be relevant to
other studies in the growing field of interpretable deep learning.
Introduction Results & Discussion
Methodology Conclusion
6/3/2023 5:46 PM 26
KPNNs Infer a Regulatory Model of T-
Cell Receptor Stimulation
Introduction Results & Discussion
Methodology Conclusion
Figure 12: TCR stimulation model as inferred by KPNN
6/3/2023 5:46 PM 27
Conclusion
Introduction Pre-processing Results & Discussion
Methodology Conclusion
• Several technologies converged to enable interpretable deep learning on
biological networks with single-cell RNA-seq data as input.
large databases of signaling pathways and gene-regulatory interactions, built on high-
throughput experiments and on manual curation of many studies.
Advancements in single-cell sequencing makes it possible to obtain transcriptome
profiles for deep learning to play out its strengths.
Deep learning as a method for inferring hidden states in deep neural networks based
on large training datasets, allows to infer unobserved states in complex biological
networks.
• By combining collection of potential regulatory interactions from public
databases with single-cell RNA-seq data for the investigated biological
system, deep neural networks may indeed be able to unravel those
regulatory mechanisms that are relevant to the biological question of
interest.
Introduction Results & Discussion
Methodology Conclusion
Reference Paper
6/3/2023 5:46 PM 28
Introduction Sample Preparation Data Analysis
Considerations References
6/3/2023 5:46 PM 29

Mais conteúdo relacionado

Semelhante a ANN in System Biology

Implementation of energy efficient coverage aware routing protocol for wirele...
Implementation of energy efficient coverage aware routing protocol for wirele...Implementation of energy efficient coverage aware routing protocol for wirele...
Implementation of energy efficient coverage aware routing protocol for wirele...ijfcstjournal
 
Functional genomics
Functional genomicsFunctional genomics
Functional genomicsPawan Kumar
 
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...Artificial Neural Network and Multi-Response Optimization in Reliability Meas...
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...inventionjournals
 
Density based spatial clustering of applications with noises for dna methylat...
Density based spatial clustering of applications with noises for dna methylat...Density based spatial clustering of applications with noises for dna methylat...
Density based spatial clustering of applications with noises for dna methylat...Atef Alghzzy
 
19 16112 paper 097 ijeecs(edit)
19 16112 paper 097 ijeecs(edit)19 16112 paper 097 ijeecs(edit)
19 16112 paper 097 ijeecs(edit)IAESIJEECS
 
Convolutional neural networks deepa
Convolutional neural networks deepaConvolutional neural networks deepa
Convolutional neural networks deepadeepa4466
 
Investigations on Hybrid Learning in ANFIS
Investigations on Hybrid Learning in ANFISInvestigations on Hybrid Learning in ANFIS
Investigations on Hybrid Learning in ANFISIJERA Editor
 
MAMMOGRAPHY LESION DETECTION USING FASTER R-CNN DETECTOR
MAMMOGRAPHY LESION DETECTION USING FASTER R-CNN DETECTORMAMMOGRAPHY LESION DETECTION USING FASTER R-CNN DETECTOR
MAMMOGRAPHY LESION DETECTION USING FASTER R-CNN DETECTORcscpconf
 
Isolated words recognition using mfcc, lpc and neural network
Isolated words recognition using mfcc, lpc and neural networkIsolated words recognition using mfcc, lpc and neural network
Isolated words recognition using mfcc, lpc and neural networkeSAT Journals
 
Black-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX modelBlack-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX modelIJECEIAES
 
Data integration lab_meeting
Data integration lab_meetingData integration lab_meeting
Data integration lab_meetingLiangqun Lu
 
Signal Processing IEEE 2015 Projects
Signal Processing IEEE 2015 ProjectsSignal Processing IEEE 2015 Projects
Signal Processing IEEE 2015 ProjectsVijay Karan
 
Lung Cancer Detection using Decision Tree Algorithm
Lung Cancer Detection using Decision Tree AlgorithmLung Cancer Detection using Decision Tree Algorithm
Lung Cancer Detection using Decision Tree AlgorithmIRJET Journal
 
International Journal of Computational Engineering Research(IJCER)
 International Journal of Computational Engineering Research(IJCER)  International Journal of Computational Engineering Research(IJCER)
International Journal of Computational Engineering Research(IJCER) ijceronline
 

Semelhante a ANN in System Biology (20)

Soft Computing.pptx
Soft Computing.pptxSoft Computing.pptx
Soft Computing.pptx
 
Prediction of pKa from chemical structure using free and open source tools
Prediction of pKa from chemical structure using free and open source toolsPrediction of pKa from chemical structure using free and open source tools
Prediction of pKa from chemical structure using free and open source tools
 
May 15 workshop
May 15  workshopMay 15  workshop
May 15 workshop
 
Implementation of energy efficient coverage aware routing protocol for wirele...
Implementation of energy efficient coverage aware routing protocol for wirele...Implementation of energy efficient coverage aware routing protocol for wirele...
Implementation of energy efficient coverage aware routing protocol for wirele...
 
Functional genomics
Functional genomicsFunctional genomics
Functional genomics
 
2224d_final
2224d_final2224d_final
2224d_final
 
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...Artificial Neural Network and Multi-Response Optimization in Reliability Meas...
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...
 
Density based spatial clustering of applications with noises for dna methylat...
Density based spatial clustering of applications with noises for dna methylat...Density based spatial clustering of applications with noises for dna methylat...
Density based spatial clustering of applications with noises for dna methylat...
 
F017533540
F017533540F017533540
F017533540
 
19 16112 paper 097 ijeecs(edit)
19 16112 paper 097 ijeecs(edit)19 16112 paper 097 ijeecs(edit)
19 16112 paper 097 ijeecs(edit)
 
Convolutional neural networks deepa
Convolutional neural networks deepaConvolutional neural networks deepa
Convolutional neural networks deepa
 
Investigations on Hybrid Learning in ANFIS
Investigations on Hybrid Learning in ANFISInvestigations on Hybrid Learning in ANFIS
Investigations on Hybrid Learning in ANFIS
 
MAMMOGRAPHY LESION DETECTION USING FASTER R-CNN DETECTOR
MAMMOGRAPHY LESION DETECTION USING FASTER R-CNN DETECTORMAMMOGRAPHY LESION DETECTION USING FASTER R-CNN DETECTOR
MAMMOGRAPHY LESION DETECTION USING FASTER R-CNN DETECTOR
 
Isolated words recognition using mfcc, lpc and neural network
Isolated words recognition using mfcc, lpc and neural networkIsolated words recognition using mfcc, lpc and neural network
Isolated words recognition using mfcc, lpc and neural network
 
1207.2600
1207.26001207.2600
1207.2600
 
Black-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX modelBlack-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX model
 
Data integration lab_meeting
Data integration lab_meetingData integration lab_meeting
Data integration lab_meeting
 
Signal Processing IEEE 2015 Projects
Signal Processing IEEE 2015 ProjectsSignal Processing IEEE 2015 Projects
Signal Processing IEEE 2015 Projects
 
Lung Cancer Detection using Decision Tree Algorithm
Lung Cancer Detection using Decision Tree AlgorithmLung Cancer Detection using Decision Tree Algorithm
Lung Cancer Detection using Decision Tree Algorithm
 
International Journal of Computational Engineering Research(IJCER)
 International Journal of Computational Engineering Research(IJCER)  International Journal of Computational Engineering Research(IJCER)
International Journal of Computational Engineering Research(IJCER)
 

Último

GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)Areesha Ahmad
 
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...Monika Rani
 
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.Nitya salvi
 
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...Sérgio Sacani
 
GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...
GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...
GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...Lokesh Kothari
 
All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...
All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...
All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...Sérgio Sacani
 
Hubble Asteroid Hunter III. Physical properties of newly found asteroids
Hubble Asteroid Hunter III. Physical properties of newly found asteroidsHubble Asteroid Hunter III. Physical properties of newly found asteroids
Hubble Asteroid Hunter III. Physical properties of newly found asteroidsSérgio Sacani
 
Botany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdfBotany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdfSumit Kumar yadav
 
Chemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfChemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfSumit Kumar yadav
 
PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...
PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...
PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...Sérgio Sacani
 
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...Lokesh Kothari
 
GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)Areesha Ahmad
 
Seismic Method Estimate velocity from seismic data.pptx
Seismic Method Estimate velocity from seismic  data.pptxSeismic Method Estimate velocity from seismic  data.pptx
Seismic Method Estimate velocity from seismic data.pptxAlMamun560346
 
Botany krishna series 2nd semester Only Mcq type questions
Botany krishna series 2nd semester Only Mcq type questionsBotany krishna series 2nd semester Only Mcq type questions
Botany krishna series 2nd semester Only Mcq type questionsSumit Kumar yadav
 
Disentangling the origin of chemical differences using GHOST
Disentangling the origin of chemical differences using GHOSTDisentangling the origin of chemical differences using GHOST
Disentangling the origin of chemical differences using GHOSTSérgio Sacani
 
Recombination DNA Technology (Nucleic Acid Hybridization )
Recombination DNA Technology (Nucleic Acid Hybridization )Recombination DNA Technology (Nucleic Acid Hybridization )
Recombination DNA Technology (Nucleic Acid Hybridization )aarthirajkumar25
 
Isotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on IoIsotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on IoSérgio Sacani
 
Biogenic Sulfur Gases as Biosignatures on Temperate Sub-Neptune Waterworlds
Biogenic Sulfur Gases as Biosignatures on Temperate Sub-Neptune WaterworldsBiogenic Sulfur Gases as Biosignatures on Temperate Sub-Neptune Waterworlds
Biogenic Sulfur Gases as Biosignatures on Temperate Sub-Neptune WaterworldsSérgio Sacani
 
Pulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceuticsPulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceuticssakshisoni2385
 

Último (20)

GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)
 
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
Vip profile Call Girls In Lonavala 9748763073 For Genuine Sex Service At Just...
 
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
 
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
 
GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...
GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...
GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...
 
All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...
All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...
All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...
 
Hubble Asteroid Hunter III. Physical properties of newly found asteroids
Hubble Asteroid Hunter III. Physical properties of newly found asteroidsHubble Asteroid Hunter III. Physical properties of newly found asteroids
Hubble Asteroid Hunter III. Physical properties of newly found asteroids
 
Botany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdfBotany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdf
 
CELL -Structural and Functional unit of life.pdf
CELL -Structural and Functional unit of life.pdfCELL -Structural and Functional unit of life.pdf
CELL -Structural and Functional unit of life.pdf
 
Chemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfChemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdf
 
PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...
PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...
PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...
 
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
 
GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)
 
Seismic Method Estimate velocity from seismic data.pptx
Seismic Method Estimate velocity from seismic  data.pptxSeismic Method Estimate velocity from seismic  data.pptx
Seismic Method Estimate velocity from seismic data.pptx
 
Botany krishna series 2nd semester Only Mcq type questions
Botany krishna series 2nd semester Only Mcq type questionsBotany krishna series 2nd semester Only Mcq type questions
Botany krishna series 2nd semester Only Mcq type questions
 
Disentangling the origin of chemical differences using GHOST
Disentangling the origin of chemical differences using GHOSTDisentangling the origin of chemical differences using GHOST
Disentangling the origin of chemical differences using GHOST
 
Recombination DNA Technology (Nucleic Acid Hybridization )
Recombination DNA Technology (Nucleic Acid Hybridization )Recombination DNA Technology (Nucleic Acid Hybridization )
Recombination DNA Technology (Nucleic Acid Hybridization )
 
Isotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on IoIsotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on Io
 
Biogenic Sulfur Gases as Biosignatures on Temperate Sub-Neptune Waterworlds
Biogenic Sulfur Gases as Biosignatures on Temperate Sub-Neptune WaterworldsBiogenic Sulfur Gases as Biosignatures on Temperate Sub-Neptune Waterworlds
Biogenic Sulfur Gases as Biosignatures on Temperate Sub-Neptune Waterworlds
 
Pulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceuticsPulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceutics
 

ANN in System Biology

  • 1. Knowledge-Primed Neural Networks Enable Biologically Interpretable Deep Learning on Single-Cell Sequencing Data Course: Machine Learning and Drug Design Nikolaus Fortelny and Chistoph Bock
  • 2. Knowledge-Primed Neural Networks (KPNNs) 6/3/2023 5:46 PM 2 Trained ANNs typically lack interpretability, i.e., the ability to provide human-understandable, high-level explanations of how they transform inputs (prediction attributes) into outputs which is a major limitation to the wider application of deep learning in biology. Figure 1: ANN vs. knowledge-primed neural networks (KPNNs) Introduction Results & Discussion Methodology Conclusion
  • 3. Methodology 6/3/2023 5:46 PM 3 Figure 2: Pipeline for the construction of KPNN Construction of general regulatory neural networks for human cell Construction of application specific KNNs for deep learning Applying Deep Learning Approach to KNNs Optimization of Learning Method to Enhance KPNN interpretability Training of the optimized KPNN using single cell RNA seq data Comparing ANN and KPNN Introduction Results & Discussion Methodology Conclusion
  • 4. 6/3/2023 5:46 PM 4 1.Constructing Generic KPNNs • Integrates multiple signaling pathways and gene-regulatory interactions into a single network. • For the biological basis, large-scale, genome-wide datasets of transcription factor target genes and protein signaling interactions from several public databases were integrated into a broad network model of the potential regulatory space in human cells. • This general regulatory network model is represented as a directed graph. Introduction Results & Discussion Methodology Conclusion
  • 5. 6/3/2023 5:46 PM 5 1.Constructing Generic KPNNs Introduction Results & Discussion Methodology Conclusion Figure 3: Generic acyclic graphs
  • 6. 6/3/2023 5:46 PM 6 2. Construction of Application-Specific KPNNs for Deep Learning • The resulting graphs are used for interpretable deep learning: Measured gene expression values (input nodes) predict the regulatory importance of connected transcription factors (hidden nodes), which predict the regulatory importance of their connected signaling proteins (hidden nodes), which then predict the regulatory importance and activation states of surface receptors, which finally predict the output (phenotype) of the specific dataset.  Two KPNNs were constructed in this study: 1. TCR signaling (TCR KPNN), to predict TCR stimulation from gene expression 2. Generalized KPNN (GEN KPNN) to predict cell states from single-cell RNA-seq datasets independent of specific receptors. Introduction Results & Discussion Methodology Conclusion
  • 7. 6/3/2023 5:46 PM 7 2. Construction of Application-Specific KPNNs for Deep Learning G1 G2 G3 TF1 TF2 TF3 TF4 SP1 SP 2 SP 3 SP 4 0 1 Objective Database Used TF- Target Gene Pair Harmonizome & TTRUST Signaling Proteins – Target Proteins SIGNOR Introduction Results & Discussion Methodology Conclusion Figure 4: Application specific KPNNs Table 1:Databases used for dataset acquisition
  • 8. 6/3/2023 5:46 PM 8 3. Datasets Used for Training KPNN Interpretable deep learning on KPNNs was developed and evaluated using simulated transcriptome profiles.  It was validated and benchmarked on the biological data of TCR stimulation. Model was further validated on 4 benchmark datasets 1. HCA 2. LCH 3. AML 4. Glioblastoma Introduction Results & Discussion Methodology Conclusion
  • 9. 3. Datasets Used for Training KPNN 6/3/2023 5:46 PM 9 • TCR Dataset downloaded from GEO (GSE137554). Class Values: TCR stimulated (1) – TCR un-stimulated samples (0) • Simulated Dataset We averaged expression counts for the unstimulated state for each gene to generate a baseline expression profile.  a positive or negative twofold change was introduced in selected genes resulting in a second (differential) gene expression profile. Class Values: TCR stimulated (1) – TCR un-stimulated samples (0). Introduction Results & Discussion Methodology Conclusion
  • 10. 4. Applying Deep learning to KPNNs 6/3/2023 5:46 PM 10 1. Preprocessing the input data 2. Training of the KPNN or ANN deep neural networks 3. Prediction evaluation Introduction Results & Discussion Methodology Conclusion Figure 5: Methodology of generic deep learning applied to KPNNs
  • 11. 4.1. Preprocessing of Input 6/3/2023 5:46 PM 11 • Input data were split into training set (60%), validation set (20%), and test set (20%) using numpy random.choice. • Gene expression data were converted to log (TPM + 1) values and normalized at the scale of 0 -1. • Normalization factors were calculated based on the training data and then applied to the validation and test data. Introduction Results & Discussion Methodology Conclusion
  • 12. 4.2. Network Training 6/3/2023 5:46 PM 12 Parameters Configurations Implementation Tensor flow Edge weights Randomly initialized Activation function Sigmoid Loss Function A weighed cross entropy with L2 regularization. Weighted class entropy- to improve training in the presence of class imbalance, the cross- entropy of each sample was weighted by the number of samples of each class. Class Weight (1/Nclasses)/(Nx/Nsamples) where Nclasses is the number of classes, Nx is the number of samples of class x, and Nsamples is the total number of samples Introduction Results & Discussion Methodology Conclusion Table 2: Configurations for training KPNNs
  • 13. 4.2. Network Training 6/3/2023 5:46 PM 13 • Error rate: True label probability – Predicted class probability • Training was terminated with early stopping, which was triggered by a specific number (“patience”) of epochs without considerable learning processes (“failed epochs”). • Epochs were considered as “failed epochs” if either (i) loss does not show significant improvement and remains relatively constant over a certain number of epochs. (ii) validation set error increased (indicative of overfitting) • Models were saved during training if they achieved considerable learning progress compared to the previously saved model. Introduction Results & Discussion Methodology Conclusion
  • 14. 5. Optimization of Learning Method to Enhance KPNN interpretability 6/3/2023 5:46 PM 14 • Three modifications were implemented to enhance biological interpretability of the trained KPNNs: dropout on hidden nodes to improve robustness. dropout on input nodes to increase quantitative interpretability. training on simulated control inputs to normalize for uneven connectivity in biological networks. Introduction Results & Discussion Methodology Conclusion
  • 15. 6.Training the KPNN on the Single Cell RNA Seq. Dataset 6/3/2023 5:46 PM 15 • Raw read counts of all 1735 cells, class values (stimulated or unstimulated), and an edge list encoding the TCR KPNN structure were provided as input. • Learning was stopped after 20 failed epochs. • During learning, new models were saved if they reduced the validation set error of the latest stored (i.e., so far best-performing) model by at least 20%. • Dropout rates from 0 to 50% were evaluated in steps of 10 percentage points. • Network training was repeated 90 times for each level of dropout Introduction Results & Discussion Methodology Conclusion
  • 16. Results and Discussion Maria Hassan Khan 6/3/2023 5:46 PM 16 Introduction Results & Discussion Methodology Conclusion
  • 17. KPNNs Establish Deep Learning on Biological Networks 6/3/2023 5:46 PM 17 • Constructed KPNNs reflected the typical flow of information in cells: Signals are transduced from receptors via signaling proteins to transcription factors, which in turn induce changes in gene expression. Introduction Results & Discussion Methodology Conclusion Figure 6: Mapping genome-wide regulatory networks to KPNNs
  • 18. KPNNs Establish Deep Learning on Biological Networks 6/3/2023 5:46 PM 18 • KPNNs can be trained based on single-cell RNA-seq data for cells in different states (e.g., subtypes of cancer and immune cells) or under different environmental exposures (e.g., receptor-stimulated vs unstimulated cells) • After successful completion of the learning process, each fitted edge weight is taken as an indicator for the relevance of the corresponding regulatory relationship in the investigated biological system. • We then derive node weights from the edge weights, in order to identify signaling proteins and transcription factors that are likely relevant in this biological system. • KPNNs thus exploit the ability of deep learning algorithms to assign meaningful weights across multiple hidden layers in KPNNs, thereby identifying and prioritizing relevant regulatory proteins for experimental validation and biological interpretation. Introduction Results & Discussion Methodology Conclusion
  • 19. TCR Signaling • KPNNs predicted TCR stimulation with a median receiver operating characteristic (ROC) area under curve (AUC) value of 0.984 whereas, ANN achieved value 0.948. 6/3/2023 5:46 PM 19 Introduction Results & Discussion Methodology Conclusion Figure 7: Schematic diagram of TCR KPNN
  • 20. 6/3/2023 5:46 PM 20 KPNN Leading to Interpretable Deep Learning Introduction Results & Discussion Methodology Conclusion Figure 8: Comparison of the number of layers in TCR KPNN vs. TCR ANN
  • 21. 6/3/2023 5:46 PM 21 KPNN Leading to Interpretable Deep Learning Introduction Results & Discussion Methodology Conclusion Figure 9: Comparison of the number of degrees of different layers in TCR KPNN vs. TCR ANN
  • 22. 6/3/2023 5:46 PM 22 Interpretability of KPNN Network Connectedness Introduction Results & Discussion Methodology Conclusion Figure 10: Comparison of the number of outdegrees of different layers in TCR KPNN vs.
  • 23. 6/3/2023 5:46 PM 23 Interpretability of KPNN Network Connectedness Introduction Results & Discussion Methodology Conclusion Figure 11: Comparison of the network connectedness of TCR KPNN vs. TCR ANN
  • 24. 6/3/2023 5:46 PM 24 Enhanced KPNN Interpretability via Optimized Learning Methods • Application of generic deep learning to KPNNs instead of ANNs, yielded high prediction accuracies with three recurrent challenges in interpreting the learned models: low reproducibility in the presence of network redundancy lack of quantitative interpretability (i.e., node weights fail to quantitatively reflect the predictive power of individual nodes) uneven connectivity inherent to biological networks, which affects node weights independent of training data Introduction Results & Discussion Methodology Conclusion
  • 25. 6/3/2023 5:46 PM 25 Enhanced KPNN Interpretability via Optimized Learning Methods 1. Hidden node dropout. 2. Input dropout. 3. Correction for network structure. • These three methodological optimizations enhanced the biological interpretability of KPNNs both on simulated and on real data. • It is anticipated that similar optimizations will also be relevant to other studies in the growing field of interpretable deep learning. Introduction Results & Discussion Methodology Conclusion
  • 26. 6/3/2023 5:46 PM 26 KPNNs Infer a Regulatory Model of T- Cell Receptor Stimulation Introduction Results & Discussion Methodology Conclusion Figure 12: TCR stimulation model as inferred by KPNN
  • 27. 6/3/2023 5:46 PM 27 Conclusion Introduction Pre-processing Results & Discussion Methodology Conclusion • Several technologies converged to enable interpretable deep learning on biological networks with single-cell RNA-seq data as input. large databases of signaling pathways and gene-regulatory interactions, built on high- throughput experiments and on manual curation of many studies. Advancements in single-cell sequencing makes it possible to obtain transcriptome profiles for deep learning to play out its strengths. Deep learning as a method for inferring hidden states in deep neural networks based on large training datasets, allows to infer unobserved states in complex biological networks. • By combining collection of potential regulatory interactions from public databases with single-cell RNA-seq data for the investigated biological system, deep neural networks may indeed be able to unravel those regulatory mechanisms that are relevant to the biological question of interest. Introduction Results & Discussion Methodology Conclusion
  • 28. Reference Paper 6/3/2023 5:46 PM 28 Introduction Sample Preparation Data Analysis Considerations References

Notas do Editor

  1. Deep learning provides a powerful method for predicting cell states from gene expression profiles. Generic artificial neural networks (ANNs, top row) are “black boxes” that provide little insight into the biology that underlies a successful prediction – for two reasons: (i) hidden nodes and edges in an ANN have no biological equivalent, which makes it difficult assign a biological interpretation to the weights of a fitted ANN model, and (ii) ANNs are inherently instable, and very different networks can achieve similar prediction performance. For certain specialized tasks with ample training data, ANNs already achieve prediction performance superior to that of human experts. However, the trained ANNs typically lack interpretability, i.e., the ability to provide human-understandable, high-level explanations of how they transform inputs (prediction attributes) into outputs (predicted class values). This lack of interpretability is a major limitation to the wider application of deep learning in biology and medicine—not only because it reduces trust and confidence in using such predictions for high-stakes applications such as clinical diagnostics [17, 18], but also because it misses important opportunities for data-driven biological discovery using deep learning.
  2. Different datasets were combined by counting the number of datasets that support a connection for each transcription factor/target gene pair. To prioritize experimental over computational evidence, weights of 1.0 and 0.3 were used for experimental and computational support, respectively. For each gene, the transcription factors with the largest weighted number of datasets supporting the connection to the gene were retained. As a result of this procedure, 745 genes were targeted by more than 25 transcription factors; these genes were removed from the analysis to avoid biasing the network by including promiscuous or false-positive regulatory interactions.
  3. Human cell atlas Langerhans cell histiocytosis (LCH)  Acute myeloid leukemia (AML) 
  4. In the context of gene expression data analysis, a normalization factor is a value used to adjust gene expression measurements across samples to account for differences in total expression levels or other sources of technical variation. The normalization factor ensures that the expression levels of genes are comparable between samples, allowing for meaningful comparisons and downstream analyses. Housekeeping Gene Normalization: Housekeeping gene normalization selects a set of genes known to be stably expressed across samples as reference genes. The expression values of the reference genes are then used to calculate a scaling factor that is applied to normalize the expression values of all genes in a sample.
  5. The sigmoid function was chosen because of its similarity to “on” and “off” states in biological systems Weighted Cross-Entropy Loss with L2 regularization is a specific combination of loss functions used in certain scenarios, often in multiclass classification tasks. Here's a breakdown of each component: Weighted Cross-Entropy Loss: Cross-entropy loss is a commonly used loss function for classification problems. It measures the dissimilarity between the predicted class probabilities and the true class labels. In weighted cross-entropy, different weights are assigned to each class to account for class imbalances or the importance of certain classes. This helps in giving more emphasis to the underrepresented classes or the classes of higher importance. L2 Regularization (also known as weight decay): L2 regularization is a technique used to prevent overfitting in machine learning models. It adds a penalty term to the loss function that discourages large weights in the model. The penalty term is proportional to the square of the weights, hence the term "L2" or "ridge" regularization. By penalizing large weights, L2 regularization encourages the model to have smaller, more evenly distributed weights, which can help in reducing overfitting and improving generalization.
  6. Training with Early Stopping: Early stopping refers to the practice of monitoring the model's performance during training and stopping the training process early if certain criteria are met. It allows for terminating training when further iterations are unlikely to improve the model's performance, thus preventing overfitting and saving computational resources. Using the most recently saved model, the error on the (previously unseen) test set was assessed and receiver-operator characteristic curves were derived as the final performance metric
  7. Dropout constitutes a modification of the learning algorithm where a given percentage of nodes is randomly set to zero in each training iteration and for each sample, temporarily removing the contribution of the affected nodes. Dropout was originally developed as a regularization technique to reduce overfitting [48]. Applied to KPNNs, dropout encourages the learning algorithm to spread weights across all relevant nodes, which reduces variability of node weights across network replicates (leading to improved robustness) and balances weights across input nodes (leading to node weights that more quantitatively reflect node importance). Dropout was implemented in our Fortelny and Bock Genome Biology (2020) 21:190 Page 24 of 36 network training program using TensorFlow nn.dropout, which was applied separately to hidden nodes and to input nodes. Normalizing for the uneven connectivity of biological networks To normalize for the biological network structure and connectivity patterns of KPNNs, we trained the KPNNs not only on the actual single-cell transcriptome data, but also on artificial control inputs where all input nodes were set to values that are equally predictive of the class values. All input nodes carry equal importance in this scenario, hence the resulting node weights reflect only the inherent network structure of the KPNN. Control inputs thus quantify the effect of uneven connectivity on node weights and allow us to normalize for it. Control inputs were generated in the same way as the simulated single-cell transcriptome datasets used to develop our methodology (as described above), with one difference: while only a subset of inputs were selected as predictive in the simulated dataset, the control inputs were simulated such that all inputs were equally predictive. To that end, raw read counts were summed up across all cells corresponding to one class value (e.g., unstimulated cells in the TCR data) to generate an average transcriptome profile. A positive or negative (randomly selected with numpy random.choice) twofold change was added to all input genes to generate a second average transcriptome profile. Reads were then drawn from the two average transcriptome profiles using numpy random.binomial [120]. The number of reads drawn was based on the number of reads in each cell in the original data. For the TCR data and simulated data, the same number of cells as in the original dataset was simulated with control inputs.
  8. Specifically, nodes with only one parent were never dropped, and dropout was limited to 10% for nodes with two Fortelny and Bock Genome Biology (2020) 21:190 Page 25 of 36 parents and to 30% for nodes with three parents. Network training was repeated 90 times for each level of dropout.
  9. The expression levels of the regulated genes were used as input nodes, whose values can be measured by single-cell RNA-seq. Signaling proteins and transcription factors were modeled as hidden nodes, and we infer the corresponding node weights during the learning process. Finally, receptors constitute the output nodes of the KPNN, capturing phenotypic cell states and the ability of individual cells to interact with their environment. The learning process typically starts from a generic, fully connected, feedforward ANN—which is a directed acyclic graph organized in layers such that each node receives input from all nodes in the previous layer and sends its output to all nodes in the next layer. During the learning process, edge weights are randomly initiated and then updated iteratively based on training data, seeking to improve the accuracy with which the ANN transforms the inputs into corresponding class values. With enough training data, large multi-layer ANNs can learn highly complex relationships between prediction attributes and class values, without requiring any prior domain knowledge. However, the resulting trained ANNs lack interpretability, given that their nodes, edges, and weights do not correspond to meaningful domain concepts To overcome the lack of interpretability of deep learning on ANNs, we sought to embed relevant biological domain knowledge directly into the neural networks that are trained by deep learning (Fig. 1). To that end, we replaced the fully connected ANNs of generic deep learning with networks derived from prior knowledge of biological networks, thereby creating “knowledge-primed neural networks” (KPNNs). In KPNNs, each node corresponds to a protein or a gene, and each edge corresponds to a potential regulatory relationship that has previously been observed in any biological context and annotated in public databases
  10. Gene expression (input nodes) provides the input for transcription factors (hidden nodes), whose outputs are used by signaling proteins (hidden nodes) to predict TCR stimulation (output node).
  11. To understand how the network structure of KPNNs may contribute to interpretable deep learning, we performed a systematic network analysis comparing the TCR KPNN with corresponding ANNs. As expected, the KPNN was much sparser than ANNs with the same number of nodes. For example, the best-performing ANN had 151-fold more edges than the KPNN (Additional file 1: Fig. S1d-f). However, the most striking differences between the KPNN and the corresponding ANNs referred to the structural properties of these networks. To understand how the network structure of KPNNs may contribute to interpretable deep learning, we performed a systematic network analysis comparing the TCR KPNN with corresponding ANNs. As expected, the KPNN was much sparser than ANNs with the same number of nodes. For example, the best-performing ANN had 151-fold more edges than the KPNN (Additional file 1: Fig. S1d-f). However, the most striking differences between the KPNN and the corresponding ANNs referred to the structural properties of these networks. When we compared the KPNN to fully connected ANNs with the same number of nodes (fANN) and to sparse ANNs with the same number of edges as the KPNN (sANN, see the “Methods” section for details), the KPNN contained many deep connections (“shortcuts”), contrasting with the stric
  12. Because each node in the fully connected ANNs is connected to all nodes of both the previous and the following layer, the distance from a given node to any node of the input layer is the same, depending only on the node’s layer. In contrast, distances in the KPNN differed widely due to the presence of shortcuts that connect certain parts of the network much more directly with the input layer. When we compared the KPNN to fully connected ANNs with the same number of nodes (fANN) and to sparse ANNs with the same number of edges as the KPNN (sANN, see the “Methods” section for details), the KPNN contained many deep connections (“shortcuts”), contrasting with the strictly layered network architecture of the ANNs (Fig. 2a, Additional file 1: Fig. S2a). Because each node in the fully connected ANNs is connected to all nodes of both the previous and the following layer, the distance from a given node to any node of the input layer is the same, depending only on the node’s layer. In contrast, distances in the KPNN differed widely due to the presence of shortcuts that connect certain parts of the network much more directly with the input layer.
  13. The KPNN was further characterized by unique patterns of network connectedness. Most notably, the distribution of node outdegrees (the outdegree measures the number of edges that leave a node) resembled an exponential distribution for the KPNN, while all nodes in the ANNs were connected to the same (constant) number of other nodes. The KPNN was further characterized by unique patterns of network connectedness. Most notably, the distribution of node outdegrees (the outdegree measures the number of edges that leave a node) resembled an exponential distribution for the KPNN, while all nodes in the ANNs were connected to the same (constant) number of other nodes (Fig. 2b, Additional file 1: Fig. S2b-c). This approximately scale-free nature of the KPNN derives in part from the presence of highly connected “hubs” in biological networks [45], which link many nodes with lower connectivity. As a result of these connectivity patterns, the KPNN showed higher modularity than the ANNs, which is evident from its higher sensitivity to fragmentation upon edge removal compared to fully connected ANNs, and even to sparse ANNs (
  14. his approximately scale-free nature of the KPNN derives in part from the presence of highly connected “hubs” in biological networks [45], which link many nodes with lower connectivity. As a result of these connectivity patterns, the KPNN showed higher modularity than the ANNs, which is evident from its higher sensitivity to fragmentation upon edge removal compared to fully connected ANNs, and even to sparse ANNs (
  15. training with hidden node dropout enhances robustness of node weights in the presence of network redundancy, input node dropout distinguishes between strongly predictive and weakly predictive nodes Calculation of node weights based on control inputs allows us to normalize for the effect of uneven connectivity in biological networks. Together with the sparse structure of KPNNs and the fact that every node and every edge in the KPNNs has a corresponding biological interpretation, this learning method fosters interpretable deep learning on biological networks
  16. When we trained the TCR KPNN with our optimized learning method, based on single-cell RNA-seq data for TCR stimulated vs unstimulated Jurkat cells [49], we obtained high prediction accuracies (Fig. 4a), comparable to those obtained by generic deep learning on KPNNs and on ANNs (Additional file 1: Fig. S1b). Moreover, we indeed observed the anticipated boost in biological interpretability of the trained KPNN model (