ProtoHPE: Prototype-guided High-frequency Patch Enhancement for
Visible-Infrared Person Re-identification
- URL: http://arxiv.org/abs/2310.07552v1
- Date: Wed, 11 Oct 2023 14:54:40 GMT
- Title: ProtoHPE: Prototype-guided High-frequency Patch Enhancement for
Visible-Infrared Person Re-identification
- Authors: Guiwei Zhang and Yongfei Zhang and Zichang Tan
- Abstract summary: Cross-modal correlated high-frequency components are less affected by variations such as wavelength, pose, and background clutter than holistic images.
We propose textbfPrototype-guided textbfHigh-frequency textbfPatch textbfEnhancement (ProtoHPE) with two core designs.
- Score: 16.634909655008254
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Visible-infrared person re-identification is challenging due to the large
modality gap. To bridge the gap, most studies heavily rely on the correlation
of visible-infrared holistic person images, which may perform poorly under
severe distribution shifts. In contrast, we find that some cross-modal
correlated high-frequency components contain discriminative visual patterns and
are less affected by variations such as wavelength, pose, and background
clutter than holistic images. Therefore, we are motivated to bridge the
modality gap based on such high-frequency components, and propose
\textbf{Proto}type-guided \textbf{H}igh-frequency \textbf{P}atch
\textbf{E}nhancement (ProtoHPE) with two core designs. \textbf{First}, to
enhance the representation ability of cross-modal correlated high-frequency
components, we split patches with such components by Wavelet Transform and
exponential moving average Vision Transformer (ViT), then empower ViT to take
the split patches as auxiliary input. \textbf{Second}, to obtain semantically
compact and discriminative high-frequency representations of the same identity,
we propose Multimodal Prototypical Contrast. To be specific, it hierarchically
captures the comprehensive semantics of different modal instances, facilitating
the aggregation of high-frequency representations belonging to the same
identity. With it, ViT can capture key high-frequency components during
inference without relying on ProtoHPE, thus bringing no extra complexity.
Extensive experiments validate the effectiveness of ProtoHPE.
Related papers
- Diffusion Transformer meets Multi-level Wavelet Spectrum for Single Image Super-Resolution [15.056888813012451]
We propose a Diffusion Transformer model based on image Wavelet spectra for SR (DTWSR)<n>DTWSR incorporates the superiority of diffusion models and transformers to capture the interrelations among multiscale frequency sub-bands.<n>A dual-decoder is designed elaborately to handle the distinct variances in low-frequency and high-frequency sub-bands, without omitting their alignment in image generation.
arXiv Detail & Related papers (2025-11-03T02:56:58Z) - THAT: Token-wise High-frequency Augmentation Transformer for Hyperspectral Pansharpening [11.026691195419453]
Transformer-based methods have demonstrated strong potential in hyperspectral pansharpening.<n>Their effectiveness is often limited by redundant token representations and a lack of multi-scale feature modeling.<n>We propose the Token-wise High-frequency Augmentation Transformer (THAT), a novel framework designed to enhance hyperspectral pansharpening.
arXiv Detail & Related papers (2025-08-11T17:03:10Z) - Freqformer: Image-Demoiréing Transformer via Efficient Frequency Decomposition [83.40450475728792]
We present Freqformer, a Transformer-based framework specifically designed for image demoir'eing through targeted frequency separation.<n>Our method performs an effective frequency decomposition that explicitly splits moir'e patterns into high-frequency spatially-localized textures and low-frequency scale-robust color distortions.<n>Experiments on various demoir'eing benchmarks demonstrate that Freqformer achieves state-of-the-art performance with a compact model size.
arXiv Detail & Related papers (2025-05-25T12:23:10Z) - PAD: Phase-Amplitude Decoupling Fusion for Multi-Modal Land Cover Classification [30.563079264213112]
Phase-Amplitude Decoupling (PAD) is a frequency-aware framework that separates phase (modality-shared) and amplitude (modality-complementary) components in the Fourier domain.<n>PAD is the first to introduce explicit amplitude-phase decoupling for multi-modal fusion.<n>Our work establishes a new paradigm for physics-aware multi-modal fusion in remote sensing.
arXiv Detail & Related papers (2025-04-27T07:21:42Z) - FUSE: Label-Free Image-Event Joint Monocular Depth Estimation via Frequency-Decoupled Alignment and Degradation-Robust Fusion [63.87313550399871]
Image-event joint depth estimation methods leverage complementary modalities for robust perception, yet face challenges in generalizability.
We propose Self-supervised Transfer (PST) and FrequencyDe-coupled Fusion module (FreDF)
PST establishes cross-modal knowledge transfer through latent space alignment with image foundation models.
FreDF explicitly decouples high-frequency edge features from low-frequency structural components, resolving modality-specific frequency mismatches.
arXiv Detail & Related papers (2025-03-25T15:04:53Z) - Prototype-Driven Multi-Feature Generation for Visible-Infrared Person Re-identification [11.664820595258988]
Primary challenges in visible-infrared person re-identification arise from the differences between visible (vis) and infrared (ir) images.
Existing methods often rely on horizontal partitioning to align part-level features, which can introduce inaccuracies.
We propose a novel Prototype-Driven Multi-feature generation framework (PDM) aimed at mitigating cross-modal discrepancies.
arXiv Detail & Related papers (2024-09-09T14:12:23Z) - Frequency Domain Modality-invariant Feature Learning for
Visible-infrared Person Re-Identification [79.9402521412239]
We propose a novel Frequency Domain modality-invariant feature learning framework (FDMNet) to reduce modality discrepancy from the frequency domain perspective.
Our framework introduces two novel modules, namely the Instance-Adaptive Amplitude Filter (IAF) and the Phrase-Preserving Normalization (PPNorm)
arXiv Detail & Related papers (2024-01-03T17:11:27Z) - Unified Frequency-Assisted Transformer Framework for Detecting and
Grounding Multi-Modal Manipulation [109.1912721224697]
We present the Unified Frequency-Assisted transFormer framework, named UFAFormer, to address the DGM4 problem.
By leveraging the discrete wavelet transform, we decompose images into several frequency sub-bands, capturing rich face forgery artifacts.
Our proposed frequency encoder, incorporating intra-band and inter-band self-attentions, explicitly aggregates forgery features within and across diverse sub-bands.
arXiv Detail & Related papers (2023-09-18T11:06:42Z) - Cross-modal Orthogonal High-rank Augmentation for RGB-Event
Transformer-trackers [58.802352477207094]
We explore the great potential of a pre-trained vision Transformer (ViT) to bridge the vast distribution gap between two modalities.
We propose a mask modeling strategy that randomly masks a specific modality of some tokens to enforce the interaction between tokens from different modalities interacting proactively.
Experiments demonstrate that our plug-and-play training augmentation techniques can significantly boost state-of-the-art one-stream and two trackersstream to a large extent in terms of both tracking precision and success rate.
arXiv Detail & Related papers (2023-07-09T08:58:47Z) - Exploring Invariant Representation for Visible-Infrared Person
Re-Identification [77.06940947765406]
Cross-spectral person re-identification, which aims to associate identities to pedestrians across different spectra, faces a main challenge of the modality discrepancy.
In this paper, we address the problem from both image-level and feature-level in an end-to-end hybrid learning framework named robust feature mining network (RFM)
Experiment results on two standard cross-spectral person re-identification datasets, RegDB and SYSU-MM01, have demonstrated state-of-the-art performance.
arXiv Detail & Related papers (2023-02-02T05:24:50Z) - Diffusion Probabilistic Model Made Slim [128.2227518929644]
We introduce a customized design for slim diffusion probabilistic models (DPM) for light-weight image synthesis.
We achieve 8-18x computational complexity reduction as compared to the latent diffusion models on a series of conditional and unconditional image generation tasks.
arXiv Detail & Related papers (2022-11-27T16:27:28Z) - Unsupervised Misaligned Infrared and Visible Image Fusion via
Cross-Modality Image Generation and Registration [59.02821429555375]
We present a robust cross-modality generation-registration paradigm for unsupervised misaligned infrared and visible image fusion.
To better fuse the registered infrared images and visible images, we present a feature Interaction Fusion Module (IFM)
arXiv Detail & Related papers (2022-05-24T07:51:57Z) - TFill: Image Completion via a Transformer-Based Architecture [69.62228639870114]
We propose treating image completion as a directionless sequence-to-sequence prediction task.
We employ a restrictive CNN with small and non-overlapping RF for token representation.
In a second phase, to improve appearance consistency between visible and generated regions, a novel attention-aware layer (AAL) is introduced.
arXiv Detail & Related papers (2021-04-02T01:42:01Z) - Generalizing Face Forgery Detection with High-frequency Features [63.33397573649408]
Current CNN-based detectors tend to overfit to method-specific color textures and thus fail to generalize.
We propose to utilize the high-frequency noises for face forgery detection.
The first is the multi-scale high-frequency feature extraction module that extracts high-frequency noises at multiple scales.
The second is the residual-guided spatial attention module that guides the low-level RGB feature extractor to concentrate more on forgery traces from a new perspective.
arXiv Detail & Related papers (2021-03-23T08:19:21Z) - Are High-Frequency Components Beneficial for Training of Generative
Adversarial Networks [11.226288436817956]
Generative Adversarial Networks (GANs) have the ability to generate realistic images that are visually indistinguishable from real images.
Recent studies of the image spectrum have demonstrated that generated and real images share significant differences at high frequency.
We propose two preprocessing methods eliminating high-frequency differences in GANs training.
arXiv Detail & Related papers (2021-03-20T04:37:06Z) - SFANet: A Spectrum-aware Feature Augmentation Network for
Visible-Infrared Person Re-Identification [12.566284647658053]
We propose a novel spectrum-aware feature augementation network named SFANet for cross-modality matching problem.
Learning with grayscale-spectrum images, our model can apparently reduce modality discrepancy and detect inner structure relations.
In feature-level, we improve the conventional two-stream network through balancing the number of specific and sharable convolutional blocks.
arXiv Detail & Related papers (2021-02-24T08:57:32Z) - Multi-Scale Cascading Network with Compact Feature Learning for
RGB-Infrared Person Re-Identification [35.55895776505113]
Multi-Scale Part-Aware Cascading framework (MSPAC) is formulated by aggregating multi-scale fine-grained features from part to global.
Cross-modality correlations can thus be efficiently explored on salient features for distinctive modality-invariant feature learning.
arXiv Detail & Related papers (2020-12-12T15:39:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.