MTP: Multi-Task Pruning for Efficient Semantic Segmentation Networks
- URL: http://arxiv.org/abs/2007.08386v2
- Date: Tue, 15 Mar 2022 06:54:53 GMT
- Title: MTP: Multi-Task Pruning for Efficient Semantic Segmentation Networks
- Authors: Xinghao Chen, Yiman Zhang, Yunhe Wang
- Abstract summary: We present a multi-task channel pruning approach for semantic segmentation networks.
The importance of each convolution filter wrt the channel of an arbitrary layer will be simultaneously determined by the classification and segmentation tasks.
Experimental results on several benchmarks illustrate the superiority of the proposed algorithm over the state-of-the-art pruning methods.
- Score: 32.84644563020912
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper focuses on channel pruning for semantic segmentation networks.
Previous methods to compress and accelerate deep neural networks in the
classification task cannot be straightforwardly applied to the semantic
segmentation network that involves an implicit multi-task learning problem via
pre-training. To identify the redundancy in segmentation networks, we present a
multi-task channel pruning approach. The importance of each convolution filter
\wrt the channel of an arbitrary layer will be simultaneously determined by the
classification and segmentation tasks. In addition, we develop an alternative
scheme for optimizing importance scores of filters in the entire network.
Experimental results on several benchmarks illustrate the superiority of the
proposed algorithm over the state-of-the-art pruning methods. Notably, we can
obtain an about $2\times$ FLOPs reduction on DeepLabv3 with only an about $1\%$
mIoU drop on the PASCAL VOC 2012 dataset and an about $1.3\%$ mIoU drop on
Cityscapes dataset, respectively.
Related papers
- SGLP: A Similarity Guided Fast Layer Partition Pruning for Compressing Large Deep Models [19.479746878680707]
Layer pruning is a potent approach to reduce network size and improve computational efficiency.
We propose a Similarity Guided fast Layer Partition pruning for compressing large deep models.
Our method outperforms the state-of-the-art methods in both accuracy and computational efficiency.
arXiv Detail & Related papers (2024-10-14T04:01:08Z) - A Point-Based Approach to Efficient LiDAR Multi-Task Perception [49.91741677556553]
PAttFormer is an efficient multi-task architecture for joint semantic segmentation and object detection in point clouds.
Unlike other LiDAR-based multi-task architectures, our proposed PAttFormer does not require separate feature encoders for task-specific point cloud representations.
Our evaluations show substantial gains from multi-task learning, improving LiDAR semantic segmentation by +1.7% in mIou and 3D object detection by +1.7% in mAP.
arXiv Detail & Related papers (2024-04-19T11:24:34Z) - Performance-aware Approximation of Global Channel Pruning for Multitask
CNNs [13.356477450355547]
Global channel pruning (GCP) aims to remove a subset of channels (filters) across different layers from a deep model without hurting the performance.
We propose a Performance-Aware Global Channel Pruning (PAGCP) framework.
Experiments on several multitask datasets show that the proposed PAGCP can reduce the FLOPs and parameters by over 60% with minor performance drop.
arXiv Detail & Related papers (2023-03-21T15:15:21Z) - Dynamic Neural Network for Multi-Task Learning Searching across Diverse
Network Topologies [14.574399133024594]
We present a new MTL framework that searches for optimized structures for multiple tasks with diverse graph topologies.
We design a restricted DAG-based central network with read-in/read-out layers to build topologically diverse task-adaptive structures.
arXiv Detail & Related papers (2023-03-13T05:01:50Z) - Group Fisher Pruning for Practical Network Compression [58.25776612812883]
We present a general channel pruning approach that can be applied to various complicated structures.
We derive a unified metric based on Fisher information to evaluate the importance of a single channel and coupled channels.
Our method can be used to prune any structures including those with coupled channels.
arXiv Detail & Related papers (2021-08-02T08:21:44Z) - Manifold Regularized Dynamic Network Pruning [102.24146031250034]
This paper proposes a new paradigm that dynamically removes redundant filters by embedding the manifold information of all instances into the space of pruned networks.
The effectiveness of the proposed method is verified on several benchmarks, which shows better performance in terms of both accuracy and computational cost.
arXiv Detail & Related papers (2021-03-10T03:59:03Z) - MetricUNet: Synergistic Image- and Voxel-Level Learning for Precise CT
Prostate Segmentation via Online Sampling [66.01558025094333]
We propose a two-stage framework, with the first stage to quickly localize the prostate region and the second stage to precisely segment the prostate.
We introduce a novel online metric learning module through voxel-wise sampling in the multi-task network.
Our method can effectively learn more representative voxel-level features compared with the conventional learning methods with cross-entropy or Dice loss.
arXiv Detail & Related papers (2020-05-15T10:37:02Z) - Network Adjustment: Channel Search Guided by FLOPs Utilization Ratio [101.84651388520584]
This paper presents a new framework named network adjustment, which considers network accuracy as a function of FLOPs.
Experiments on standard image classification datasets and a wide range of base networks demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2020-04-06T15:51:00Z) - A "Network Pruning Network" Approach to Deep Model Compression [62.68120664998911]
We present a filter pruning approach for deep model compression using a multitask network.
Our approach is based on learning a a pruner network to prune a pre-trained target network.
The compressed model produced by our approach is generic and does not need any special hardware/software support.
arXiv Detail & Related papers (2020-01-15T20:38:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.