ShoeRinsics: Shoeprint Prediction for Forensics with Intrinsic
Decomposition
- URL: http://arxiv.org/abs/2205.02361v1
- Date: Wed, 4 May 2022 23:42:55 GMT
- Title: ShoeRinsics: Shoeprint Prediction for Forensics with Intrinsic
Decomposition
- Authors: Samia Shafique, Bailey Kong, Shu Kong, Charless C. Fowlkes
- Abstract summary: We propose to leverage shoe tread photographs collected by online retailers.
We develop a model that performs intrinsic image decomposition from a single tread photo.
Our approach, which we term ShoeRinsics, combines domain adaptation and re-rendering losses in order to leverage a mix of fully supervised synthetic data and unsupervised retail image data.
- Score: 29.408442567550004
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Shoe tread impressions are one of the most common types of evidence left at
crime scenes. However, the utility of such evidence is limited by the lack of
databases of footwear impression patterns that cover the huge and growing
number of distinct shoe models. We propose to address this gap by leveraging
shoe tread photographs collected by online retailers. The core challenge is to
predict the impression pattern from the shoe photograph since ground-truth
impressions or 3D shapes of tread patterns are not available. We develop a
model that performs intrinsic image decomposition (predicting depth, normal,
albedo, and lighting) from a single tread photo. Our approach, which we term
ShoeRinsics, combines domain adaptation and re-rendering losses in order to
leverage a mix of fully supervised synthetic data and unsupervised retail image
data. To validate model performance, we also collected a set of paired
shoe-sole images and corresponding prints, and define a benchmarking protocol
to quantify the accuracy of predicted impressions. On this benchmark,
ShoeRinsics outperforms existing methods for depth prediction and
synthetic-to-real domain adaptation.
Related papers
- Data Augmentation via Latent Diffusion for Saliency Prediction [67.88936624546076]
Saliency prediction models are constrained by the limited diversity and quantity of labeled data.
We propose a novel data augmentation method for deep saliency prediction that edits natural images while preserving the complexity and variability of real-world scenes.
arXiv Detail & Related papers (2024-09-11T14:36:24Z) - Adversarial Robustification via Text-to-Image Diffusion Models [56.37291240867549]
Adrial robustness has been conventionally believed as a challenging property to encode for neural networks.
We develop a scalable and model-agnostic solution to achieve adversarial robustness without using any data.
arXiv Detail & Related papers (2024-07-26T10:49:14Z) - CriSp: Leveraging Tread Depth Maps for Enhanced Crime-Scene Shoeprint Matching [8.153893958726117]
Shoeprints are a common type of evidence found at crime scenes and are used regularly in forensic investigations.
Existing methods cannot effectively employ deep learning techniques to match noisy and occluded crime-scene shoeprints to a shoe database.
We propose CriSp, which matches crime-scene shoeprints to tread depth maps by training on this data.
CriSp significantly outperforms state-of-the-art methods in both automated shoeprint matching and image retrieval tailored to this task.
arXiv Detail & Related papers (2024-04-25T18:50:26Z) - ShoeModel: Learning to Wear on the User-specified Shoes via Diffusion Model [60.60623356092564]
We propose a shoe-wearing system, called Shoe-Model, to generate plausible images of human legs interacting with the given shoes.
Compared to baselines, our ShoeModel is shown to generalize better to different type of shoes and has ability of keeping the ID-consistency of the given shoes.
arXiv Detail & Related papers (2024-04-07T06:56:51Z) - Improving and Evaluating Machine Learning Methods for Forensic Shoeprint Matching [0.2509487459755192]
We propose a machine learning pipeline for forensic shoeprint pattern matching.
We extract 2D coordinates from shoeprint scans using edge detection and align the two shoeprints with iterative closest point (ICP)
We then extract similarity metrics to quantify how well the two prints match and use these metrics to train a random forest.
arXiv Detail & Related papers (2024-04-02T15:24:25Z) - StableVITON: Learning Semantic Correspondence with Latent Diffusion
Model for Virtual Try-On [35.227896906556026]
Given a clothing image and a person image, an image-based virtual try-on aims to generate a customized image that appears natural and accurately reflects the characteristics of the clothing image.
In this work, we aim to expand the applicability of the pre-trained diffusion model so that it can be utilized independently for the virtual try-on task.
Our proposed zero cross-attention blocks not only preserve the clothing details by learning the semantic correspondence but also generate high-fidelity images by utilizing the inherent knowledge of the pre-trained model in the warping process.
arXiv Detail & Related papers (2023-12-04T08:27:59Z) - How Will It Drape Like? Capturing Fabric Mechanics from Depth Images [7.859729554664895]
We propose a method to estimate the mechanical parameters of fabrics using a casual capture setup with a depth camera.
Our approach enables to create mechanically-correct digital representations of real-world textile materials.
arXiv Detail & Related papers (2023-04-13T17:54:08Z) - Masked Images Are Counterfactual Samples for Robust Fine-tuning [77.82348472169335]
Fine-tuning deep learning models can lead to a trade-off between in-distribution (ID) performance and out-of-distribution (OOD) robustness.
We propose a novel fine-tuning method, which uses masked images as counterfactual samples that help improve the robustness of the fine-tuning model.
arXiv Detail & Related papers (2023-03-06T11:51:28Z) - Person Image Synthesis via Denoising Diffusion Model [116.34633988927429]
We show how denoising diffusion models can be applied for high-fidelity person image synthesis.
Our results on two large-scale benchmarks and a user study demonstrate the photorealism of our proposed approach under challenging scenarios.
arXiv Detail & Related papers (2022-11-22T18:59:50Z) - Learning Intrinsic Images for Clothing [10.21096394185778]
In this paper, we focus on intrinsic image decomposition for clothing images.
A more interpretable edge-aware metric and an annotation scheme is designed for the testing set.
We show that our proposed model significantly reduce texture-copying artifacts while retaining surprisingly tiny details.
arXiv Detail & Related papers (2021-11-16T14:43:12Z) - Scene Uncertainty and the Wellington Posterior of Deterministic Image
Classifiers [68.9065881270224]
We introduce the Wellington Posterior, which is the distribution of outcomes that would have been obtained in response to data that could have been generated by the same scene.
We explore the use of data augmentation, dropout, ensembling, single-view reconstruction, and model linearization to compute a Wellington Posterior.
Additional methods include the use of conditional generative models such as generative adversarial networks, neural radiance fields, and conditional prior networks.
arXiv Detail & Related papers (2021-06-25T20:10:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.