Feature Gradient Flow for Interpreting Deep Neural Networks in Head and
Neck Cancer Prediction
- URL: http://arxiv.org/abs/2307.13061v1
- Date: Mon, 24 Jul 2023 18:25:59 GMT
- Title: Feature Gradient Flow for Interpreting Deep Neural Networks in Head and
Neck Cancer Prediction
- Authors: Yinzhu Jin, Jonathan C. Garneau, P. Thomas Fletcher
- Abstract summary: This paper introduces feature gradient flow, a new technique for interpreting deep learning models in terms of features that are understandable to humans.
We measure the agreement of interpretable features with the gradient flow of a model.
We develop a technique for training neural networks to be more interpretable by adding a regularization term to the loss function.
- Score: 2.9477900773805032
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper introduces feature gradient flow, a new technique for interpreting
deep learning models in terms of features that are understandable to humans.
The gradient flow of a model locally defines nonlinear coordinates in the input
data space representing the information the model is using to make its
decisions. Our idea is to measure the agreement of interpretable features with
the gradient flow of a model. To then evaluate the importance of a particular
feature to the model, we compare that feature's gradient flow measure versus
that of a baseline noise feature. We then develop a technique for training
neural networks to be more interpretable by adding a regularization term to the
loss function that encourages the model gradients to align with those of chosen
interpretable features. We test our method in a convolutional neural network
prediction of distant metastasis of head and neck cancer from a computed
tomography dataset from the Cancer Imaging Archive.
Related papers
- Measuring Feature Dependency of Neural Networks by Collapsing Feature Dimensions in the Data Manifold [18.64569268049846]
We introduce a new technique to measure the feature dependency of neural network models.
The motivation is to better understand a model by querying whether it is using information from human-understandable features.
We test our method on deep neural network models trained on synthetic image data with known ground truth.
arXiv Detail & Related papers (2024-04-18T17:10:18Z) - Rolling bearing fault diagnosis method based on generative adversarial enhanced multi-scale convolutional neural network model [7.600902237804825]
A rolling bearing fault diagnosis method based on generative adversarial enhanced multi-scale convolutional neural network model is proposed.
Compared with ResNet method, the experimental results show that the proposed method has better generalization performance and anti-noise performance.
arXiv Detail & Related papers (2024-03-21T06:42:35Z) - Generalizing Backpropagation for Gradient-Based Interpretability [103.2998254573497]
We show that the gradient of a model is a special case of a more general formulation using semirings.
This observation allows us to generalize the backpropagation algorithm to efficiently compute other interpretable statistics.
arXiv Detail & Related papers (2023-07-06T15:19:53Z) - Deep Manifold Learning with Graph Mining [80.84145791017968]
We propose a novel graph deep model with a non-gradient decision layer for graph mining.
The proposed model has achieved state-of-the-art performance compared to the current models.
arXiv Detail & Related papers (2022-07-18T04:34:08Z) - Contrastive Brain Network Learning via Hierarchical Signed Graph Pooling
Model [64.29487107585665]
Graph representation learning techniques on brain functional networks can facilitate the discovery of novel biomarkers for clinical phenotypes and neurodegenerative diseases.
Here, we propose an interpretable hierarchical signed graph representation learning model to extract graph-level representations from brain functional networks.
In order to further improve the model performance, we also propose a new strategy to augment functional brain network data for contrastive learning.
arXiv Detail & Related papers (2022-07-14T20:03:52Z) - Using a novel fractional-order gradient method for CNN back-propagation [1.6679382332181059]
Researchers propose a novel deep learning model and apply it to COVID-19 diagnosis.
Model uses the tool of fractional calculus, which has the potential to improve the performance of gradient methods.
arXiv Detail & Related papers (2022-05-01T23:38:06Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - Self-Supervised Graph Representation Learning for Neuronal Morphologies [75.38832711445421]
We present GraphDINO, a data-driven approach to learn low-dimensional representations of 3D neuronal morphologies from unlabeled datasets.
We show, in two different species and across multiple brain areas, that this method yields morphological cell type clusterings on par with manual feature-based classification by experts.
Our method could potentially enable data-driven discovery of novel morphological features and cell types in large-scale datasets.
arXiv Detail & Related papers (2021-12-23T12:17:47Z) - Sampling and Recovery of Graph Signals based on Graph Neural Networks [44.76396026242879]
We propose interpretable graph neural networks for sampling and recovery of graph signals, respectively.
The proposed methods are able to flexibly learn a variety of graph signal models from data by leveraging the learning ability of neural networks.
In the experiments, we illustrate the effects of the proposed graph neural sampling and recovery modules and find that the modules can flexibly adapt to various graph structures and graph signals.
arXiv Detail & Related papers (2020-11-03T01:45:41Z) - A Bayesian Perspective on Training Speed and Model Selection [51.15664724311443]
We show that a measure of a model's training speed can be used to estimate its marginal likelihood.
We verify our results in model selection tasks for linear models and for the infinite-width limit of deep neural networks.
Our results suggest a promising new direction towards explaining why neural networks trained with gradient descent are biased towards functions that generalize well.
arXiv Detail & Related papers (2020-10-27T17:56:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.