Exposing and addressing the fragility of neural networks in digital
pathology
- URL: http://arxiv.org/abs/2206.15274v1
- Date: Thu, 30 Jun 2022 13:25:34 GMT
- Title: Exposing and addressing the fragility of neural networks in digital
pathology
- Authors: Joona Pohjonen, Carolin St\"urenberg, Atte F\"ohr, Esa Pitk\"anen,
Antti Rannikko, Tuomas Mirtti
- Abstract summary: textttStrongAugment is evaluated with large-scale, heterogeneous histopathology data.
neural networks trained with textttStrongAugment retain similar performance on all datasets.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Neural networks have achieved impressive results in many medical imaging
tasks but often perform substantially worse on out-of-distribution datasets
originating from different medical centres or patient cohorts. Evaluating this
lack of ability to generalise and address the underlying problem are the two
main challenges in developing neural networks intended for clinical practice.
In this study, we develop a new method for evaluating neural network models'
ability to generalise by generating a large number of distribution-shifted
datasets, which can be used to thoroughly investigate their robustness to
variability encountered in clinical practice. Compared to external validation,
\textit{shifted evaluation} can provide explanations for why neural networks
fail on a given dataset, thus offering guidance on how to improve model
robustness. With shifted evaluation, we demonstrate that neural networks,
trained with state-of-the-art methods, are highly fragile to even small
distribution shifts from training data, and in some cases lose all
discrimination ability.
To address this fragility, we develop an augmentation strategy, explicitly
designed to increase neural networks' robustness to distribution shifts.
\texttt{StrongAugment} is evaluated with large-scale, heterogeneous
histopathology data including five training datasets from two tissue types, 274
distribution-shifted datasets and 20 external datasets from four countries.
Neural networks trained with \texttt{StrongAugment} retain similar performance
on all datasets, even with distribution shifts where networks trained with
current state-of-the-art methods lose all discrimination ability. We recommend
using strong augmentation and shifted evaluation to train and evaluate all
neural networks intended for clinical practice.
Related papers
- Feature Averaging: An Implicit Bias of Gradient Descent Leading to Non-Robustness in Neural Networks [13.983863226803336]
We argue that "Feature Averaging" is one of the principal factors contributing to non-robustness of deep neural networks.
We provide a detailed theoretical analysis of the training dynamics of gradient descent in a two-layer ReLU network for a binary classification task.
We prove that, with the provision of more granular supervised information, a two-layer multi-class neural network is capable of learning individual features.
arXiv Detail & Related papers (2024-10-14T09:28:32Z) - Graph Neural Networks for Learning Equivariant Representations of Neural Networks [55.04145324152541]
We propose to represent neural networks as computational graphs of parameters.
Our approach enables a single model to encode neural computational graphs with diverse architectures.
We showcase the effectiveness of our method on a wide range of tasks, including classification and editing of implicit neural representations.
arXiv Detail & Related papers (2024-03-18T18:01:01Z) - Deep Neural Networks Tend To Extrapolate Predictably [51.303814412294514]
neural network predictions tend to be unpredictable and overconfident when faced with out-of-distribution (OOD) inputs.
We observe that neural network predictions often tend towards a constant value as input data becomes increasingly OOD.
We show how one can leverage our insights in practice to enable risk-sensitive decision-making in the presence of OOD inputs.
arXiv Detail & Related papers (2023-10-02T03:25:32Z) - Graph Neural Operators for Classification of Spatial Transcriptomics
Data [1.408706290287121]
We propose a study incorporating various graph neural network approaches to validate the efficacy of applying neural operators towards prediction of brain regions in mouse brain tissue samples.
We were able to achieve an F1 score of nearly 72% for the graph neural operator approach which outperformed all baseline and other graph network approaches.
arXiv Detail & Related papers (2023-02-01T18:32:06Z) - Global quantitative robustness of regression feed-forward neural
networks [0.0]
We adapt the notion of the regression breakdown point to regression neural networks.
We compare the performance, measured by the out-of-sample loss, by a proxy of the breakdown rate.
The results indeed motivate to use robust loss functions for neural network training.
arXiv Detail & Related papers (2022-11-18T09:57:53Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - Anomaly Detection on Attributed Networks via Contrastive Self-Supervised
Learning [50.24174211654775]
We present a novel contrastive self-supervised learning framework for anomaly detection on attributed networks.
Our framework fully exploits the local information from network data by sampling a novel type of contrastive instance pair.
A graph neural network-based contrastive learning model is proposed to learn informative embedding from high-dimensional attributes and local structure.
arXiv Detail & Related papers (2021-02-27T03:17:20Z) - Domain Generalization for Medical Imaging Classification with
Linear-Dependency Regularization [59.5104563755095]
We introduce a simple but effective approach to improve the generalization capability of deep neural networks in the field of medical imaging classification.
Motivated by the observation that the domain variability of the medical images is to some extent compact, we propose to learn a representative feature space through variational encoding.
arXiv Detail & Related papers (2020-09-27T12:30:30Z) - ResiliNet: Failure-Resilient Inference in Distributed Neural Networks [56.255913459850674]
We introduce ResiliNet, a scheme for making inference in distributed neural networks resilient to physical node failures.
Failout simulates physical node failure conditions during training using dropout, and is specifically designed to improve the resiliency of distributed neural networks.
arXiv Detail & Related papers (2020-02-18T05:58:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.