Grafted network for person re-identification
- URL: http://arxiv.org/abs/2006.01967v2
- Date: Sat, 6 Jun 2020 05:25:28 GMT
- Title: Grafted network for person re-identification
- Authors: Jiabao Wang, Yang Li, Shanshan Jiao, Zhuang Miao, Rui Zhang
- Abstract summary: Convolutional neural networks have shown outstanding effectiveness in person re-identification (re-ID)
We propose a novel grafted network (GraftedNet), which is designed by grafting a high-accuracy rootstock and a light-weighted scion.
Experimental results show that the proposed GraftedNet achieves 93.02%, 85.3% and 76.2% in Rank-1 and 81.6%, 74.7% and 71.6% in mAP, with only 4.6M parameters.
- Score: 14.372506245952383
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Convolutional neural networks have shown outstanding effectiveness in person
re-identification (re-ID). However, the models always have large number of
parameters and much computation for mobile application. In order to relieve
this problem, we propose a novel grafted network (GraftedNet), which is
designed by grafting a high-accuracy rootstock and a light-weighted scion. The
rootstock is based on the former parts of ResNet-50 to provide a strong
baseline, while the scion is a new designed module, composed of the latter
parts of SqueezeNet, to compress the parameters. To extract more discriminative
feature representation, a joint multi-level and part-based feature is proposed.
In addition, to train GraftedNet efficiently, we propose an accompanying
learning method, by adding an accompanying branch to train the model in
training and removing it in testing for saving parameters and computation. On
three public person re-ID benchmarks (Market1501, DukeMTMC-reID and CUHK03),
the effectiveness of GraftedNet are evaluated and its components are analyzed.
Experimental results show that the proposed GraftedNet achieves 93.02%, 85.3%
and 76.2% in Rank-1 and 81.6%, 74.7% and 71.6% in mAP, with only 4.6M
parameters.
Related papers
- Performance Analysis of Various EfficientNet Based U-Net++ Architecture
for Automatic Building Extraction from High Resolution Satellite Images [0.0]
Building extraction heavily relies on semantic segmentation of high-resolution remote sensing imagery.
Various efficientNet backbone based U-Net++ has been proposed in this study.
According on the experimental findings, the suggested model significantly outperforms previous cutting-edge approaches.
arXiv Detail & Related papers (2023-09-05T18:14:14Z) - Systematic Architectural Design of Scale Transformed Attention Condenser
DNNs via Multi-Scale Class Representational Response Similarity Analysis [93.0013343535411]
We propose a novel type of analysis called Multi-Scale Class Representational Response Similarity Analysis (ClassRepSim)
We show that adding STAC modules to ResNet style architectures can result in up to a 1.6% increase in top-1 accuracy.
Results from ClassRepSim analysis can be used to select an effective parameterization of the STAC module resulting in competitive performance.
arXiv Detail & Related papers (2023-06-16T18:29:26Z) - CNN-transformer mixed model for object detection [3.5897534810405403]
In this paper, I propose a convolutional module with a transformer.
It aims to improve the recognition accuracy of the model by fusing the detailed features extracted by CNN with the global features extracted by a transformer.
After 100 rounds of training on the Pascal VOC dataset, the accuracy of the results reached 81%, which is 4.6 better than the faster RCNN[4] using resnet101[5] as the backbone.
arXiv Detail & Related papers (2022-12-13T16:35:35Z) - Prompt Tuning for Parameter-efficient Medical Image Segmentation [79.09285179181225]
We propose and investigate several contributions to achieve a parameter-efficient but effective adaptation for semantic segmentation on two medical imaging datasets.
We pre-train this architecture with a dedicated dense self-supervision scheme based on assignments to online generated prototypes.
We demonstrate that the resulting neural network model is able to attenuate the gap between fully fine-tuned and parameter-efficiently adapted models.
arXiv Detail & Related papers (2022-11-16T21:55:05Z) - Learning to Learn with Generative Models of Neural Network Checkpoints [71.06722933442956]
We construct a dataset of neural network checkpoints and train a generative model on the parameters.
We find that our approach successfully generates parameters for a wide range of loss prompts.
We apply our method to different neural network architectures and tasks in supervised and reinforcement learning.
arXiv Detail & Related papers (2022-09-26T17:59:58Z) - EdgeNeXt: Efficiently Amalgamated CNN-Transformer Architecture for
Mobile Vision Applications [68.35683849098105]
We introduce split depth-wise transpose attention (SDTA) encoder that splits input tensors into multiple channel groups.
Our EdgeNeXt model with 1.3M parameters achieves 71.2% top-1 accuracy on ImageNet-1K.
Our EdgeNeXt model with 5.6M parameters achieves 79.4% top-1 accuracy on ImageNet-1K.
arXiv Detail & Related papers (2022-06-21T17:59:56Z) - LilNetX: Lightweight Networks with EXtreme Model Compression and
Structured Sparsification [36.651329027209634]
LilNetX is an end-to-end trainable technique for neural networks.
It enables learning models with specified accuracy-rate-computation trade-off.
arXiv Detail & Related papers (2022-04-06T17:59:10Z) - Bayesian Convolutional Neural Networks for Seven Basic Facial Expression
Classifications [5.365808418695478]
Seven basic facial expression classifications are a basic way to express complex human emotions.
Based on the traditional Bayesian neural network framework, the ResNet18_BNN network constructed in this paper has been improved.
arXiv Detail & Related papers (2021-07-10T13:09:13Z) - Non-Parametric Adaptive Network Pruning [125.4414216272874]
We introduce non-parametric modeling to simplify the algorithm design.
Inspired by the face recognition community, we use a message passing algorithm to obtain an adaptive number of exemplars.
EPruner breaks the dependency on the training data in determining the "important" filters.
arXiv Detail & Related papers (2021-01-20T06:18:38Z) - Inception Convolution with Efficient Dilation Search [121.41030859447487]
Dilation convolution is a critical mutant of standard convolution neural network to control effective receptive fields and handle large scale variance of objects.
We propose a new mutant of dilated convolution, namely inception (dilated) convolution where the convolutions have independent dilation among different axes, channels and layers.
We explore a practical method for fitting the complex inception convolution to the data, a simple while effective dilation search algorithm(EDO) based on statistical optimization is developed.
arXiv Detail & Related papers (2020-12-25T14:58:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.