Beta-Rank: A Robust Convolutional Filter Pruning Method For Imbalanced
Medical Image Analysis
- URL: http://arxiv.org/abs/2304.07461v2
- Date: Mon, 26 Jun 2023 03:35:51 GMT
- Title: Beta-Rank: A Robust Convolutional Filter Pruning Method For Imbalanced
Medical Image Analysis
- Authors: Morteza Homayounfar, Mohamad Koohi-Moghadam, Reza Rawassizadeh, Varut
Vardhanabhuti
- Abstract summary: Deep neural networks include a high number of parameters and operations.
It can be a challenge to implement these models on devices with limited computational resources.
We propose a novel filter pruning method by considering the input and output of filters along with the values of the filters that deal with imbalanced datasets better than others.
- Score: 1.3443196224057659
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: As deep neural networks include a high number of parameters and operations,
it can be a challenge to implement these models on devices with limited
computational resources. Despite the development of novel pruning methods
toward resource-efficient models, it has become evident that these models are
not capable of handling "imbalanced" and "limited number of data points". We
proposed a novel filter pruning method by considering the input and output of
filters along with the values of the filters that deal with imbalanced datasets
better than others. Our pruning method considers the fact that all information
about the importance of a filter may not be reflected in the value of the
filter. Instead, it is reflected in the changes made to the data after the
filter is applied to it. In this work, three methods are compared with the same
training conditions except for the ranking values of each method, and 14
methods are compared from other papers. We demonstrated that our model
performed significantly better than other methods for imbalanced medical
datasets. For example, when we removed up to 58% of FLOPs for the IDRID dataset
and up to 45% for the ISIC dataset, our model was able to yield an equivalent
(or even superior) result to the baseline model. To evaluate FLOP and parameter
reduction using our model in real-world settings, we built a smartphone app,
where we demonstrated a reduction of up to 79% in memory usage and 72% in
prediction time. All codes and parameters for training different models are
available at https://github.com/mohofar/Beta-Rank
Related papers
- When to Trust Your Data: Enhancing Dyna-Style Model-Based Reinforcement Learning With Data Filter [7.886307329450978]
Dyna-style algorithms combine two approaches by using simulated data from an estimated environmental model to accelerate model-free training.
Previous works address this issue by using model ensembles or pretraining the estimated model with data collected from the real environment.
We introduce an out-of-distribution data filter that removes simulated data from the estimated model that significantly diverges from data collected in the real environment.
arXiv Detail & Related papers (2024-10-16T01:49:03Z) - ScalingFilter: Assessing Data Quality through Inverse Utilization of Scaling Laws [67.59263833387536]
ScalingFilter is a novel approach that evaluates text quality based on the perplexity difference between two language models trained on the same data.
To assess the bias introduced by quality filtering, we introduce semantic diversity, a metric of utilizing text embedding models for semantic representations.
arXiv Detail & Related papers (2024-08-15T17:59:30Z) - CoLoR-Filter: Conditional Loss Reduction Filtering for Targeted Language Model Pre-training [10.511388205893295]
We propose a data selection method, CoLoR-Filter, which leverages an empirical Bayes-inspired approach to derive a simple and computationally efficient selection criterion.
CoLoR-Filter can train a 1.2b parameter target model to match a 1.2b parameter model trained on 25b randomly selected tokens with 25x less data for Books and 11x less data for the downstream tasks.
arXiv Detail & Related papers (2024-06-15T15:28:02Z) - From Data Deluge to Data Curation: A Filtering-WoRA Paradigm for Efficient Text-based Person Search [19.070305201045954]
In text-based person search endeavors, data generation has emerged as a prevailing practice, addressing concerns over privacy preservation and the arduous task of manual annotation.
We observe that only a subset of the data in constructed datasets plays a decisive role.
We introduce a new Filtering-WoRA paradigm, which contains a filtering algorithm to identify this crucial data subset and WoRA learning strategy for light fine-tuning.
arXiv Detail & Related papers (2024-04-16T05:29:14Z) - Filter Pruning based on Information Capacity and Independence [11.411996979581295]
This paper introduces a new filter pruning method that selects filters in an interpretable, multi-perspective, and lightweight manner.
For the amount of information contained in each filter, a new metric called information capacity is proposed.
For correlations among filters, another metric called information independence is designed.
arXiv Detail & Related papers (2023-03-07T04:26:44Z) - Knockoffs-SPR: Clean Sample Selection in Learning with Noisy Labels [56.81761908354718]
We propose a novel theoretically guaranteed clean sample selection framework for learning with noisy labels.
Knockoffs-SPR can be regarded as a sample selection module for a standard supervised training pipeline.
We further combine it with a semi-supervised algorithm to exploit the support of noisy data as unlabeled data.
arXiv Detail & Related papers (2023-01-02T07:13:28Z) - Training Compact CNNs for Image Classification using Dynamic-coded
Filter Fusion [139.71852076031962]
We present a novel filter pruning method, dubbed dynamic-coded filter fusion (DCFF)
We derive compact CNNs in a computation-economical and regularization-free manner for efficient image classification.
Our DCFF derives a compact VGGNet-16 with only 72.77M FLOPs and 1.06M parameters while reaching top-1 accuracy of 93.47%.
arXiv Detail & Related papers (2021-07-14T18:07:38Z) - Deep Model Compression based on the Training History [13.916984628784768]
We propose a novel History Based Filter Pruning (HBFP) method that utilizes network training history for filter pruning.
The proposed pruning method outperforms the state-of-the-art in terms of FLOPs reduction (floating-point operations) by 97.98%, 83.42%, 78.43%, and 74.95% for LeNet-5, VGG-16, ResNet-56, and ResNet-110 models, respectively.
arXiv Detail & Related papers (2021-01-30T06:04:21Z) - Non-Parametric Adaptive Network Pruning [125.4414216272874]
We introduce non-parametric modeling to simplify the algorithm design.
Inspired by the face recognition community, we use a message passing algorithm to obtain an adaptive number of exemplars.
EPruner breaks the dependency on the training data in determining the "important" filters.
arXiv Detail & Related papers (2021-01-20T06:18:38Z) - Data Agnostic Filter Gating for Efficient Deep Networks [72.4615632234314]
Current filter pruning methods mainly leverage feature maps to generate important scores for filters and prune those with smaller scores.
In this paper, we propose a data filter pruning method that uses an auxiliary network named Dagger module to induce pruning.
In addition, to help prune filters with certain FLOPs constraints, we leverage an explicit FLOPs-aware regularization to directly promote pruning filters toward target FLOPs.
arXiv Detail & Related papers (2020-10-28T15:26:40Z) - Filter Grafting for Deep Neural Networks [71.39169475500324]
Filter grafting aims to improve the representation capability of Deep Neural Networks (DNNs)
We develop an entropy-based criterion to measure the information of filters and an adaptive weighting strategy for balancing the grafted information among networks.
For example, the grafted MobileNetV2 outperforms the non-grafted MobileNetV2 by about 7 percent on CIFAR-100 dataset.
arXiv Detail & Related papers (2020-01-15T03:18:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.