Search By Image: Deeply Exploring Beneficial Features for Beauty Product
Retrieval
- URL: http://arxiv.org/abs/2303.14075v1
- Date: Fri, 24 Mar 2023 15:38:58 GMT
- Title: Search By Image: Deeply Exploring Beneficial Features for Beauty Product
Retrieval
- Authors: Mingqiang Wei, Qian Sun, Haoran Xie, Dong Liang, Fu Lee Wang
- Abstract summary: This paper studies a practically meaningful problem of beauty product retrieval (BPR) by neural networks.
We broadly extract different types of image features, and raise an intriguing question that whether these features are beneficial to i) suppress data variations of real-world captured images.
We present a novel variable-attention neural network to understand the combination of multiple features (termed VM-Net) of beauty product images.
- Score: 21.78262478923889
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Searching by image is popular yet still challenging due to the extensive
interference arose from i) data variations (e.g., background, pose, visual
angle, brightness) of real-world captured images and ii) similar images in the
query dataset. This paper studies a practically meaningful problem of beauty
product retrieval (BPR) by neural networks. We broadly extract different types
of image features, and raise an intriguing question that whether these features
are beneficial to i) suppress data variations of real-world captured images,
and ii) distinguish one image from others which look very similar but are
intrinsically different beauty products in the dataset, therefore leading to an
enhanced capability of BPR. To answer it, we present a novel variable-attention
neural network to understand the combination of multiple features (termed
VM-Net) of beauty product images. Considering that there are few publicly
released training datasets for BPR, we establish a new dataset with more than
one million images classified into more than 20K categories to improve both the
generalization and anti-interference abilities of VM-Net and other methods. We
verify the performance of VM-Net and its competitors on the benchmark dataset
Perfect-500K, where VM-Net shows clear improvements over the competitors in
terms of MAP@7. The source code and dataset will be released upon publication.
Related papers
- Rethinking Image Super-Resolution from Training Data Perspectives [54.28824316574355]
We investigate the understudied effect of the training data used for image super-resolution (SR)
With this, we propose an automated image evaluation pipeline.
We find that datasets with (i) low compression artifacts, (ii) high within-image diversity as judged by the number of different objects, and (iii) a large number of images from ImageNet or PASS all positively affect SR performance.
arXiv Detail & Related papers (2024-09-01T16:25:04Z) - ImagiNet: A Multi-Content Dataset for Generalizable Synthetic Image Detection via Contrastive Learning [0.0]
Generative models produce images with a level of authenticity nearly indistinguishable from real photos and artwork.
The difficulty of identifying synthetic images leaves online media platforms vulnerable to impersonation and misinformation attempts.
We introduce ImagiNet, a high-resolution and balanced dataset for synthetic image detection.
arXiv Detail & Related papers (2024-07-29T13:57:24Z) - Improving Human-Object Interaction Detection via Virtual Image Learning [68.56682347374422]
Human-Object Interaction (HOI) detection aims to understand the interactions between humans and objects.
In this paper, we propose to alleviate the impact of such an unbalanced distribution via Virtual Image Leaning (VIL)
A novel label-to-image approach, Multiple Steps Image Creation (MUSIC), is proposed to create a high-quality dataset that has a consistent distribution with real images.
arXiv Detail & Related papers (2023-08-04T10:28:48Z) - JourneyDB: A Benchmark for Generative Image Understanding [89.02046606392382]
We introduce a comprehensive dataset, referred to as JourneyDB, that caters to the domain of generative images.
Our meticulously curated dataset comprises 4 million distinct and high-quality generated images.
On our dataset, we have devised four benchmarks to assess the performance of generated image comprehension.
arXiv Detail & Related papers (2023-07-03T02:39:08Z) - Improving Image Recognition by Retrieving from Web-Scale Image-Text Data [68.63453336523318]
We introduce an attention-based memory module, which learns the importance of each retrieved example from the memory.
Compared to existing approaches, our method removes the influence of the irrelevant retrieved examples, and retains those that are beneficial to the input query.
We show that it achieves state-of-the-art accuracies in ImageNet-LT, Places-LT and Webvision datasets.
arXiv Detail & Related papers (2023-04-11T12:12:05Z) - Towards Robustness of Neural Networks [0.0]
We introduce ImageNet-A/O and ImageNet-R as well as a synthetic environment and testing suite we called CAOS.
All of the datasets were created for testing robustness and measuring progress in robustness.
We build off of simple baselines in the form of Maximum Logit, and Typicality Score as well as create a novel data augmentation method in the form of DeepAugment.
arXiv Detail & Related papers (2021-12-30T19:41:10Z) - Detection and Localization of Multiple Image Splicing Using MobileNet V1 [0.0]
Two or more images are combined to generate a new image that can transmit information across social media platforms.
This research work proposes multiple image splicing forgery detection using Mask R-CNN, with a backbone as a MobileNet V1.
It also calculates the percentage score of a forged region of multiple spliced images.
arXiv Detail & Related papers (2021-08-22T09:27:22Z) - AugNet: End-to-End Unsupervised Visual Representation Learning with
Image Augmentation [3.6790362352712873]
We propose AugNet, a new deep learning training paradigm to learn image features from a collection of unlabeled pictures.
Our experiments demonstrate that the method is able to represent the image in low dimensional space.
Unlike many deep-learning-based image retrieval algorithms, our approach does not require access to external annotated datasets.
arXiv Detail & Related papers (2021-06-11T09:02:30Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z) - Real-MFF: A Large Realistic Multi-focus Image Dataset with Ground Truth [58.226535803985804]
We introduce a large and realistic multi-focus dataset called Real-MFF.
The dataset contains 710 pairs of source images with corresponding ground truth images.
We evaluate 10 typical multi-focus algorithms on this dataset for the purpose of illustration.
arXiv Detail & Related papers (2020-03-28T12:33:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.