Guided Hybrid Quantization for Object detection in Multimodal Remote
Sensing Imagery via One-to-one Self-teaching
- URL: http://arxiv.org/abs/2301.00131v1
- Date: Sat, 31 Dec 2022 06:14:59 GMT
- Title: Guided Hybrid Quantization for Object detection in Multimodal Remote
Sensing Imagery via One-to-one Self-teaching
- Authors: Jiaqing Zhang, Jie Lei, Weiying Xie, Yunsong Li, Xiuping Jia
- Abstract summary: We propose a Guided Hybrid Quantization with One-to-one Self-Teaching (GHOST) framework.
First, we first design a structure called guided quantization self-distillation (GQSD), which is an innovative idea for realizing lightweight through the synergy of quantization and distillation.
Third, in order to improve information transformation, we propose a one-to-one self-teaching (OST) module to give the student network a ability of self-judgment.
- Score: 35.316067181895264
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Considering the computation complexity, we propose a Guided Hybrid
Quantization with One-to-one Self-Teaching (GHOST}) framework. More concretely,
we first design a structure called guided quantization self-distillation
(GQSD), which is an innovative idea for realizing lightweight through the
synergy of quantization and distillation. The training process of the
quantization model is guided by its full-precision model, which is time-saving
and cost-saving without preparing a huge pre-trained model in advance. Second,
we put forward a hybrid quantization (HQ) module to obtain the optimal bit
width automatically under a constrained condition where a threshold for
distribution distance between the center and samples is applied in the weight
value search space. Third, in order to improve information transformation, we
propose a one-to-one self-teaching (OST) module to give the student network a
ability of self-judgment. A switch control machine (SCM) builds a bridge
between the student network and teacher network in the same location to help
the teacher to reduce wrong guidance and impart vital knowledge to the student.
This distillation method allows a model to learn from itself and gain
substantial improvement without any additional supervision. Extensive
experiments on a multimodal dataset (VEDAI) and single-modality datasets (DOTA,
NWPU, and DIOR) show that object detection based on GHOST outperforms the
existing detectors. The tiny parameters (<9.7 MB) and Bit-Operations (BOPs)
(<2158 G) compared with any remote sensing-based, lightweight or
distillation-based algorithms demonstrate the superiority in the lightweight
design domain. Our code and model will be released at
https://github.com/icey-zhang/GHOST.
Related papers
- Intra-class Patch Swap for Self-Distillation [3.282914142012984]
We propose a teacher-free distillation framework based on a single student network.<n>Our approach is built on a simple yet highly effective augmentation, called intra-class patch swap augmentation.<n>Our method consistently outperforms both existing self-distillation baselines and conventional teacher-based KD approaches.
arXiv Detail & Related papers (2025-05-20T09:30:19Z) - Multi-modal Knowledge Distillation-based Human Trajectory Forecasting [35.060041571520024]
Pedestrian trajectory forecasting is crucial in various applications such as autonomous driving and mobile robot navigation.
In such applications, camera-based perception enables the extraction of additional modalities (human pose, text) to enhance prediction accuracy.
We propose a multi-modal knowledge distillation framework: a student model with limited modality is distilled from a teacher model trained with full range of modalities.
arXiv Detail & Related papers (2025-03-28T07:32:51Z) - Promoting CNNs with Cross-Architecture Knowledge Distillation for Efficient Monocular Depth Estimation [4.242540533823568]
Transformer models are usually computationally-expensive, and their effectiveness in light-weight models are limited compared to convolutions.
We propose a cross-architecture knowledge distillation method for MDE, dubbed DisDepth, to enhance efficient CNN models with the supervision of state-of-the-art transformer models.
Our method achieves significant improvements on various efficient backbones, showcasing its potential for efficient monocular depth estimation.
arXiv Detail & Related papers (2024-04-25T07:55:47Z) - Model-aware reinforcement learning for high-performance Bayesian
experimental design in quantum metrology [0.5461938536945721]
Quantum sensors offer control flexibility during estimation by allowing manipulation by the experimenter across various parameters.
We introduce a versatile procedure capable of optimizing a wide range of problems in quantum metrology, estimation, and hypothesis testing.
We combine model-aware reinforcement learning (RL) with Bayesian estimation based on particle filtering.
arXiv Detail & Related papers (2023-12-28T12:04:15Z) - One-Step Diffusion Distillation via Deep Equilibrium Models [64.11782639697883]
We introduce a simple yet effective means of distilling diffusion models directly from initial noise to the resulting image.
Our method enables fully offline training with just noise/image pairs from the diffusion model.
We demonstrate that the DEQ architecture is crucial to this capability, as GET matches a $5times$ larger ViT in terms of FID scores.
arXiv Detail & Related papers (2023-12-12T07:28:40Z) - Complexity Matters: Rethinking the Latent Space for Generative Modeling [65.64763873078114]
In generative modeling, numerous successful approaches leverage a low-dimensional latent space, e.g., Stable Diffusion.
In this study, we aim to shed light on this under-explored topic by rethinking the latent space from the perspective of model complexity.
arXiv Detail & Related papers (2023-07-17T07:12:29Z) - Disentanglement via Latent Quantization [60.37109712033694]
In this work, we construct an inductive bias towards encoding to and decoding from an organized latent space.
We demonstrate the broad applicability of this approach by adding it to both basic data-re (vanilla autoencoder) and latent-reconstructing (InfoGAN) generative models.
arXiv Detail & Related papers (2023-05-28T06:30:29Z) - Vertical Layering of Quantized Neural Networks for Heterogeneous
Inference [57.42762335081385]
We study a new vertical-layered representation of neural network weights for encapsulating all quantized models into a single one.
We can theoretically achieve any precision network for on-demand service while only needing to train and maintain one model.
arXiv Detail & Related papers (2022-12-10T15:57:38Z) - Semi-supervised 3D Object Detection with Proficient Teachers [114.54835359657707]
Dominated point cloud-based 3D object detectors in autonomous driving scenarios rely heavily on the huge amount of accurately labeled samples.
Pseudo-Labeling methodology is commonly used for SSL frameworks, however, the low-quality predictions from the teacher model have seriously limited its performance.
We propose a new Pseudo-Labeling framework for semi-supervised 3D object detection, by enhancing the teacher model to a proficient one with several necessary designs.
arXiv Detail & Related papers (2022-07-26T04:54:03Z) - Boosting 3D Object Detection by Simulating Multimodality on Point Clouds [51.87740119160152]
This paper presents a new approach to boost a single-modality (LiDAR) 3D object detector by teaching it to simulate features and responses that follow a multi-modality (LiDAR-image) detector.
The approach needs LiDAR-image data only when training the single-modality detector, and once well-trained, it only needs LiDAR data at inference.
Experimental results on the nuScenes dataset show that our approach outperforms all SOTA LiDAR-only 3D detectors.
arXiv Detail & Related papers (2022-06-30T01:44:30Z) - HyperTransformer: Model Generation for Supervised and Semi-Supervised
Few-Shot Learning [14.412066456583917]
We propose a transformer-based model for few-shot learning that generates weights of a convolutional neural network (CNN) directly from support samples.
Our method is particularly effective for small target CNN architectures where learning a fixed universal task-independent embedding is not optimal.
We extend our approach to a semi-supervised regime utilizing unlabeled samples in the support set and further improving few-shot performance.
arXiv Detail & Related papers (2022-01-11T20:15:35Z) - Transfer learning of phase transitions in percolation and directed
percolation [2.0342076109301583]
We apply domain adversarial neural network (DANN) based on transfer learning to studying non-equilibrium and equilibrium phase transition models.
The DANN learning of both models yields reliable results which are comparable to the ones from Monte Carlo simulations.
arXiv Detail & Related papers (2021-12-31T15:24:09Z) - Distilling Object Detectors with Task Adaptive Regularization [97.52935611385179]
Current state-of-the-art object detectors are at the expense of high computational costs and are hard to deploy to low-end devices.
Knowledge distillation, which aims at training a smaller student network by transferring knowledge from a larger teacher model, is one of the promising solutions for model miniaturization.
arXiv Detail & Related papers (2020-06-23T15:58:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.