Finding the Reflection Point: Unpadding Images to Remove Data Augmentation Artifacts in Large Open Source Image Datasets for Machine Learning
- URL: http://arxiv.org/abs/2504.03168v1
- Date: Fri, 04 Apr 2025 04:54:10 GMT
- Title: Finding the Reflection Point: Unpadding Images to Remove Data Augmentation Artifacts in Large Open Source Image Datasets for Machine Learning
- Authors: Lucas Choi, Ross Greer,
- Abstract summary: We propose a systematic algorithm to delineate the reflection boundary through a minimum mean squared error approach.<n>Our method effectively identifies the transition between authentic content and its mirrored counterpart, even in the presence of compression or noise.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we address a novel image restoration problem relevant to machine learning dataset curation: the detection and removal of noisy mirrored padding artifacts. While data augmentation techniques like padding are necessary for standardizing image dimensions, they can introduce artifacts that degrade model evaluation when datasets are repurposed across domains. We propose a systematic algorithm to precisely delineate the reflection boundary through a minimum mean squared error approach with thresholding and remove reflective padding. Our method effectively identifies the transition between authentic content and its mirrored counterpart, even in the presence of compression or interpolation noise. We demonstrate our algorithm's efficacy on the SHEL5k dataset, showing significant performance improvements in zero-shot object detection tasks using OWLv2, with average precision increasing from 0.47 to 0.61 for hard hat detection and from 0.68 to 0.73 for person detection. By addressing annotation inconsistencies and distorted objects in padded regions, our approach enhances dataset integrity, enabling more reliable model evaluation across computer vision tasks.
Related papers
- Utilizing Multi-step Loss for Single Image Reflection Removal [0.9208007322096532]
Distorted images can negatively impact tasks like object detection and image segmentation.<n>We present a novel approach for image reflection removal using a single image.
arXiv Detail & Related papers (2024-12-11T17:57:25Z) - Understanding and Improving Training-Free AI-Generated Image Detections with Vision Foundation Models [68.90917438865078]
Deepfake techniques for facial synthesis and editing pose serious risks for generative models.<n>In this paper, we investigate how detection performance varies across model backbones, types, and datasets.<n>We introduce Contrastive Blur, which enhances performance on facial images, and MINDER, which addresses noise type bias, balancing performance across domains.
arXiv Detail & Related papers (2024-11-28T13:04:45Z) - Overcoming Scene Context Constraints for Object Detection in wild using Defilters [3.038642416291856]
High-level computer vision tasks such as object detection, recognition, and segmentation are particularly sensitive to image distortion.
We propose an image defilter to rectify image distortion prior to object detection.
This method enhances object detection accuracy, as models perform optimally when trained on non-distorted images.
arXiv Detail & Related papers (2024-04-12T07:30:52Z) - DVMNet++: Rethinking Relative Pose Estimation for Unseen Objects [59.51874686414509]
Existing approaches typically predict 3D translation utilizing the ground-truth object bounding box and approximate 3D rotation with a large number of discrete hypotheses.<n>We present a Deep Voxel Matching Network (DVMNet++) that computes the relative object pose in a single pass.<n>Our approach delivers more accurate relative pose estimates for novel objects at a lower computational cost compared to state-of-the-art methods.
arXiv Detail & Related papers (2024-03-20T15:41:32Z) - A Semi-supervised Object Detection Algorithm for Underwater Imagery [10.017195276758455]
We propose to treat artificial objects as anomalies and detect them through a semi-supervised framework based on Variational Autoencoders (VAEs)
We develop a method which clusters image data in a learned low-dimensional latent space and extracts images that are likely to contain anomalous features.
We demonstrate that by applying both methods on large image datasets, human operators can be shown candidate anomalous samples with a low false positive rate to identify objects of interest.
arXiv Detail & Related papers (2023-06-07T23:40:04Z) - Exploiting the Distortion-Semantic Interaction in Fisheye Data [12.633032175875865]
Fisheye data has the wider field of view advantage over other types of cameras, but this comes at the expense of high distortion.
objects further from the center exhibit deformations that make it difficult for a model to identify their semantic context.
We introduce an approach to exploit this relationship by first extracting distortion class labels based on an object's distance from the center of the image.
We then shape a backbone's representation space with a weighted contrastive loss that constrains objects of the same semantic class and distortion class to be close to each other.
arXiv Detail & Related papers (2023-04-28T20:23:38Z) - Benchmarking performance of object detection under image distortions in
an uncontrolled environment [0.483420384410068]
robustness of object detection algorithms plays a prominent role in real-world applications.
It has been proven that the performance of object detection methods suffers from in-capture distortions.
We present a performance evaluation framework for the state-of-the-art object detection methods.
arXiv Detail & Related papers (2022-10-28T09:06:52Z) - High-resolution Iterative Feedback Network for Camouflaged Object
Detection [128.893782016078]
Spotting camouflaged objects that are visually assimilated into the background is tricky for object detection algorithms.
We aim to extract the high-resolution texture details to avoid the detail degradation that causes blurred vision in edges and boundaries.
We introduce a novel HitNet to refine the low-resolution representations by high-resolution features in an iterative feedback manner.
arXiv Detail & Related papers (2022-03-22T11:20:21Z) - Contemplating real-world object classification [53.10151901863263]
We reanalyze the ObjectNet dataset recently proposed by Barbu et al. containing objects in daily life situations.
We find that applying deep models to the isolated objects, rather than the entire scene as is done in the original paper, results in around 20-30% performance improvement.
arXiv Detail & Related papers (2021-03-08T23:29:59Z) - Data Augmentation for Object Detection via Differentiable Neural
Rendering [71.00447761415388]
It is challenging to train a robust object detector when annotated data is scarce.
Existing approaches to tackle this problem include semi-supervised learning that interpolates labeled data from unlabeled data.
We introduce an offline data augmentation method for object detection, which semantically interpolates the training data with novel views.
arXiv Detail & Related papers (2021-03-04T06:31:06Z) - DetCo: Unsupervised Contrastive Learning for Object Detection [64.22416613061888]
Unsupervised contrastive learning achieves great success in learning image representations with CNN.
We present a novel contrastive learning approach, named DetCo, which fully explores the contrasts between global image and local image patches.
DetCo consistently outperforms supervised method by 1.6/1.2/1.0 AP on Mask RCNN-C4/FPN/RetinaNet with 1x schedule.
arXiv Detail & Related papers (2021-02-09T12:47:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.