Target Aware Network Architecture Search and Compression for Efficient
Knowledge Transfer
- URL: http://arxiv.org/abs/2205.05967v2
- Date: Wed, 24 Jan 2024 12:00:47 GMT
- Title: Target Aware Network Architecture Search and Compression for Efficient
Knowledge Transfer
- Authors: S.H.Shabbeer Basha, Debapriya Tula, Sravan Kumar Vinakota, Shiv Ram
Dubey
- Abstract summary: We propose a two-stage framework called TASCNet which enables efficient knowledge transfer.
TASCNet reduces the computational complexity of pre-trained CNNs over the target task by reducing both trainable parameters and FLOPs.
Similar to computer vision tasks, we have also conducted experiments on Movie Review Sentiment Analysis task.
- Score: 9.434523476406424
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Transfer Learning enables Convolutional Neural Networks (CNN) to acquire
knowledge from a source domain and transfer it to a target domain, where
collecting large-scale annotated examples is time-consuming and expensive.
Conventionally, while transferring the knowledge learned from one task to
another task, the deeper layers of a pre-trained CNN are finetuned over the
target dataset. However, these layers are originally designed for the source
task which may be over-parameterized for the target task. Thus, finetuning
these layers over the target dataset may affect the generalization ability of
the CNN due to high network complexity. To tackle this problem, we propose a
two-stage framework called TASCNet which enables efficient knowledge transfer.
In the first stage, the configuration of the deeper layers is learned
automatically and finetuned over the target dataset. Later, in the second
stage, the redundant filters are pruned from the fine-tuned CNN to decrease the
network's complexity for the target task while preserving the performance. This
two-stage mechanism finds a compact version of the pre-trained CNN with optimal
structure (number of filters in a convolutional layer, number of neurons in a
dense layer, and so on) from the hypothesis space. The efficacy of the proposed
method is evaluated using VGG-16, ResNet-50, and DenseNet-121 on CalTech-101,
CalTech-256, and Stanford Dogs datasets. Similar to computer vision tasks, we
have also conducted experiments on Movie Review Sentiment Analysis task. The
proposed TASCNet reduces the computational complexity of pre-trained CNNs over
the target task by reducing both trainable parameters and FLOPs which enables
resource-efficient knowledge transfer. The source code is available at:
https://github.com/Debapriya-Tula/TASCNet.
Related papers
- Enhancing Small Object Encoding in Deep Neural Networks: Introducing
Fast&Focused-Net with Volume-wise Dot Product Layer [0.0]
We introduce Fast&Focused-Net, a novel deep neural network architecture tailored for encoding small objects into fixed-length feature vectors.
Fast&Focused-Net employs a series of our newly proposed layer, the Volume-wise Dot Product (VDP) layer, designed to address several inherent limitations of CNNs.
For small object classification tasks, our network outperformed state-of-the-art methods on datasets such as CIFAR-10, CIFAR-100, STL-10, SVHN-Cropped, and Fashion-MNIST.
In the context of larger image classification, when combined with a transformer encoder (ViT
arXiv Detail & Related papers (2024-01-18T09:31:25Z) - Transferability of Convolutional Neural Networks in Stationary Learning
Tasks [96.00428692404354]
We introduce a novel framework for efficient training of convolutional neural networks (CNNs) for large-scale spatial problems.
We show that a CNN trained on small windows of such signals achieves a nearly performance on much larger windows without retraining.
Our results show that the CNN is able to tackle problems with many hundreds of agents after being trained with fewer than ten.
arXiv Detail & Related papers (2023-07-21T13:51:45Z) - Optimal transfer protocol by incremental layer defrosting [66.76153955485584]
Transfer learning is a powerful tool enabling model training with limited amounts of data.
The simplest transfer learning protocol is based on freezing" the feature-extractor layers of a network pre-trained on a data-rich source task.
We show that this protocol is often sub-optimal and the largest performance gain may be achieved when smaller portions of the pre-trained network are kept frozen.
arXiv Detail & Related papers (2023-03-02T17:32:11Z) - Speeding Up EfficientNet: Selecting Update Blocks of Convolutional
Neural Networks using Genetic Algorithm in Transfer Learning [0.0]
We devise a genetic algorithm to select blocks of layers for updating the parameters.
We show that our algorithm yields similar or better results than the baseline in terms of accuracy.
We also devise a metric called block importance to measure efficacy of each block as update block.
arXiv Detail & Related papers (2023-03-01T06:35:29Z) - SAR Despeckling Using Overcomplete Convolutional Networks [53.99620005035804]
despeckling is an important problem in remote sensing as speckle degrades SAR images.
Recent studies show that convolutional neural networks(CNNs) outperform classical despeckling methods.
This study employs an overcomplete CNN architecture to focus on learning low-level features by restricting the receptive field.
We show that the proposed network improves despeckling performance compared to recent despeckling methods on synthetic and real SAR images.
arXiv Detail & Related papers (2022-05-31T15:55:37Z) - An Experimental Study of the Impact of Pre-training on the Pruning of a
Convolutional Neural Network [0.0]
In recent years, deep neural networks have known a wide success in various application domains.
Deep neural networks usually involve a large number of parameters, which correspond to the weights of the network.
The pruning methods notably attempt to reduce the size of the parameter set, by identifying and removing the irrelevant weights.
arXiv Detail & Related papers (2021-12-15T16:02:15Z) - RIFLE: Backpropagation in Depth for Deep Transfer Learning through
Re-Initializing the Fully-connected LayEr [60.07531696857743]
Fine-tuning the deep convolution neural network(CNN) using a pre-trained model helps transfer knowledge learned from larger datasets to the target task.
We propose RIFLE - a strategy that deepens backpropagation in transfer learning settings.
RIFLE brings meaningful updates to the weights of deep CNN layers and improves low-level feature learning.
arXiv Detail & Related papers (2020-07-07T11:27:43Z) - AutoTune: Automatically Tuning Convolutional Neural Networks for
Improved Transfer Learning [13.909484906513102]
We introduce a mechanism for automatically tuning the Convolutional Neural Networks (CNN) for improved transfer learning.
The pre-trained CNN layers are tuned with the knowledge from target data using Bayesian Optimization.
Experiments are conducted on three benchmark datasets, e.g., CalTech-101, CalTech-256, and Stanford Dogs.
arXiv Detail & Related papers (2020-04-25T10:42:06Z) - Network Adjustment: Channel Search Guided by FLOPs Utilization Ratio [101.84651388520584]
This paper presents a new framework named network adjustment, which considers network accuracy as a function of FLOPs.
Experiments on standard image classification datasets and a wide range of base networks demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2020-04-06T15:51:00Z) - Curriculum By Smoothing [52.08553521577014]
Convolutional Neural Networks (CNNs) have shown impressive performance in computer vision tasks such as image classification, detection, and segmentation.
We propose an elegant curriculum based scheme that smoothes the feature embedding of a CNN using anti-aliasing or low-pass filters.
As the amount of information in the feature maps increases during training, the network is able to progressively learn better representations of the data.
arXiv Detail & Related papers (2020-03-03T07:27:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.