BGM: Background Mixup for X-ray Prohibited Items Detection
- URL: http://arxiv.org/abs/2412.00460v3
- Date: Fri, 25 Jul 2025 03:07:59 GMT
- Title: BGM: Background Mixup for X-ray Prohibited Items Detection
- Authors: Weizhe Liu, Renshuai Tao, Hongguang Zhu, Yunda Sun, Yao Zhao, Yunchao Wei,
- Abstract summary: Background Mixup (BGM) is a background-based augmentation technique tailored for X-ray security imaging domain.<n>Unlike conventional methods, BGM is founded on an in-depth analysis of physical properties.<n>BGM mixes background patches across regions on both 1) texture structure and 2) material variation, to benefit models from complicated background cues.
- Score: 75.58709178012502
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Current data-driven approaches for X-ray prohibited items detection remain under-explored, particularly in the design of effective data augmentations. Existing natural image augmentations for reflected light imaging neglect the data characteristics of X-ray security images. Moreover, prior X-ray augmentation methods have predominantly focused on foreground prohibited items, overlooking informative background cues. In this paper, we propose Background Mixup (BGM), a background-based augmentation technique tailored for X-ray security imaging domain. Unlike conventional methods, BGM is founded on an in-depth analysis of physical properties including: 1) X-ray Transmission Imagery: Transmitted X-ray pixels represent composite information from multiple materials along the imaging path. 2) Material-based Pseudo-coloring: Pseudo-coloring in X-ray images correlates directly with material properties, aiding in material distinction. Building upon the above insights, BGM mixes background patches across regions on both 1) texture structure and 2) material variation, to benefit models from complicated background cues. This enhances the model's capability to handle domain-specific challenges such as occlusion-induced discriminative imbalance. Importantly, BGM is orthogonal and fully compatible with existing foreground-focused augmentation techniques, enabling joint use to further enhance detection performance. Extensive experiments on multiple X-ray security benchmarks show that BGM consistently surpasses strong baselines, without additional annotations or significant training overhead. This work pioneers the exploration of background-aware augmentation in X-ray prohibited items detection and provides a lightweight, plug-and-play solution with broad applicability.
Related papers
- Superpowering Open-Vocabulary Object Detectors for X-ray Vision [53.07098133237041]
Open-vocabulary object detection (OvOD) is set to revolutionize security screening by enabling systems to recognize any item in X-ray scans.
We propose RAXO, a framework that repurposes off-the-shelf RGB OvOD detectors for robust X-ray detection.
RAXO builds high-quality X-ray class descriptors using a dual-source retrieval strategy.
arXiv Detail & Related papers (2025-03-21T11:54:16Z) - X-Field: A Physically Grounded Representation for 3D X-ray Reconstruction [25.13707706037451]
X-ray imaging is indispensable in medical diagnostics, yet its use is tightly regulated due to potential health risks.
Recent research focuses on generating novel views from sparse inputs and reconstructing Computed Tomography (CT) volumes.
We introduce X-Field, the first 3D representation specifically designed for X-ray imaging.
arXiv Detail & Related papers (2025-03-11T16:31:56Z) - Augmentation Matters: A Mix-Paste Method for X-Ray Prohibited Item Detection under Noisy Annotations [52.065764858163476]
Automatic X-ray prohibited item detection is vital for public safety.
Existing deep learning-based methods all assume that the annotations of training X-ray images are correct.
We propose an effective label-aware mixed patch paste augmentation method (Mix-Paste)
We show the superiority of our method on X-ray datasets under noisy annotations.
arXiv Detail & Related papers (2025-01-03T09:51:51Z) - Understanding and Improving Training-Free AI-Generated Image Detections with Vision Foundation Models [68.90917438865078]
Deepfake techniques for facial synthesis and editing pose serious risks for generative models.<n>In this paper, we investigate how detection performance varies across model backbones, types, and datasets.<n>We introduce Contrastive Blur, which enhances performance on facial images, and MINDER, which addresses noise type bias, balancing performance across domains.
arXiv Detail & Related papers (2024-11-28T13:04:45Z) - Dual-view X-ray Detection: Can AI Detect Prohibited Items from Dual-view X-ray Images like Humans? [78.26435264182763]
We introduce the Large-scale Dual-view X-ray (LDXray), which consists of 353,646 instances across 12 categories.<n>To emulate human intelligence in dual-view detection, we propose the Auxiliary-view Enhanced Network (AENet)<n>Experiments on the LDXray dataset demonstrate that the dual-view mechanism significantly enhances detection performance.
arXiv Detail & Related papers (2024-11-27T06:36:20Z) - Enhancing Prohibited Item Detection through X-ray-Specific Augmentation and Contextual Feature Integration [81.11400642272976]
X-ray prohibited item detection faces challenges due to the long-tail distribution and unique characteristics of X-ray imaging.
Traditional data augmentation strategies, such as copy-paste and mixup, are ineffective at improving the detection of rare items.
We propose the X-ray Imaging-driven Detection Network (XIDNet) to address these challenges.
arXiv Detail & Related papers (2024-11-27T06:13:56Z) - HF-Fed: Hierarchical based customized Federated Learning Framework for X-Ray Imaging [0.0]
In clinical applications, X-ray technology is vital for noninvasive examinations like mammography, providing essential anatomical information.
X-ray reconstruction is crucial in medical imaging for detailed visual representations of internal structures, aiding diagnosis and treatment without invasive procedures.
Recent advancements in deep learning have shown promise in X-ray reconstruction, but conventional DL methods often require centralized aggregation of large datasets.
We introduce the Hierarchical Framework-based Federated Learning method (HF-Fed) for customized X-ray imaging.
arXiv Detail & Related papers (2024-07-25T05:21:48Z) - Parents and Children: Distinguishing Multimodal DeepFakes from Natural Images [60.34381768479834]
Recent advancements in diffusion models have enabled the generation of realistic deepfakes from textual prompts in natural language.
We pioneer a systematic study on deepfake detection generated by state-of-the-art diffusion models.
arXiv Detail & Related papers (2023-04-02T10:25:09Z) - Employing similarity to highlight differences: On the impact of
anatomical assumptions in chest X-ray registration methods [2.080328156648695]
We develop an anatomically penalized convolutional multi-stage solution on the National Institutes of Health (NIH) data set.
Our method proves to be a natural way to limit the folding percentage of the warp field to 1/6 of the state of the art.
We statistically evaluate the benefits of our method and highlight the limits of currently used metrics for registration of chest X-rays.
arXiv Detail & Related papers (2023-01-23T09:42:49Z) - On the impact of using X-ray energy response imagery for object
detection via Convolutional Neural Networks [17.639472693362926]
We study the impact of variant X-ray imagery, i.e. X-ray energy response (high, low) and effective-z compared to geometries.
We evaluate CNN architectures to explore the transferability of models trained with such 'raw' variant imagery.
arXiv Detail & Related papers (2021-08-27T21:28:28Z) - Cross-Modal Contrastive Learning for Abnormality Classification and
Localization in Chest X-rays with Radiomics using a Feedback Loop [63.81818077092879]
We propose an end-to-end semi-supervised cross-modal contrastive learning framework for medical images.
We first apply an image encoder to classify the chest X-rays and to generate the image features.
The radiomic features are then passed through another dedicated encoder to act as the positive sample for the image features generated from the same chest X-ray.
arXiv Detail & Related papers (2021-04-11T09:16:29Z) - Image Separation with Side Information: A Connected Auto-Encoders Based
Approach [18.18248997032482]
We deal with the problem of separating mixed X-ray images originating from the radiography of double-sided paintings.
We propose a new Neural Network architecture, based upon 'connected' auto-encoders, designed to separate the mixed X-ray image into two simulated X-ray images corresponding to each side.
These tests show that the proposed approach outperforms other state-of-the-art X-ray image separation methods for art investigation applications.
arXiv Detail & Related papers (2020-09-16T18:39:42Z) - Occluded Prohibited Items Detection: an X-ray Security Inspection
Benchmark and De-occlusion Attention Module [50.75589128518707]
We contribute the first high-quality object detection dataset for security inspection, named OPIXray.
OPIXray focused on the widely-occurred prohibited item "cutter", annotated manually by professional inspectors from the international airport.
We propose the De-occlusion Attention Module (DOAM), a plug-and-play module that can be easily inserted into and thus promote most popular detectors.
arXiv Detail & Related papers (2020-04-18T16:10:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.