OPA: Object Placement Assessment Dataset
- URL: http://arxiv.org/abs/2107.01889v1
- Date: Mon, 5 Jul 2021 09:23:53 GMT
- Title: OPA: Object Placement Assessment Dataset
- Authors: Liu Liu, Bo Zhang, Jiangtong Li, Li Niu, Qingyang Liu, Liqing Zhang
- Abstract summary: Image composition aims to generate realistic composite image by inserting an object from one image into another background image.
In this paper, we focus on object placement assessment task, which verifies whether a composite image is plausible in terms of the object placement.
- Score: 20.791187775546625
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Image composition aims to generate realistic composite image by inserting an
object from one image into another background image, where the placement (e.g.,
location, size, occlusion) of inserted object may be unreasonable, which would
significantly degrade the quality of the composite image. Although some works
attempted to learn object placement to create realistic composite images, they
did not focus on assessing the plausibility of object placement. In this paper,
we focus on object placement assessment task, which verifies whether a
composite image is plausible in terms of the object placement. To accomplish
this task, we construct the first Object Placement Assessment (OPA) dataset
consisting of composite images and their rationality labels. Dataset is
available at https://github.com/bcmi/Object-Placement-Assessment-Dataset-OPA.
Related papers
- Generative Location Modeling for Spatially Aware Object Insertion [35.62317512925592]
Generative models have become a powerful tool for image editing tasks, including object insertion.
In this paper, we focus on creating a location model dedicated to identifying realistic object locations.
Specifically, we train an autoregressive model that generates bounding box coordinates, conditioned on the background image and the desired object class.
This formulation allows to effectively handle sparse placement annotations and to incorporate implausible locations into a preference dataset by performing direct preference optimization.
arXiv Detail & Related papers (2024-10-17T14:00:41Z) - DESOBAv2: Towards Large-scale Real-world Dataset for Shadow Generation [19.376935979734714]
In this work, we focus on generating plausible shadow for the inserted foreground object to make the composite image more realistic.
To supplement the existing small-scale dataset DESOBA, we create a large-scale dataset called DESOBAv2.
arXiv Detail & Related papers (2023-08-19T10:21:23Z) - Foreground Object Search by Distilling Composite Image Feature [15.771802337102837]
Foreground object search (FOS) aims to find compatible foreground objects for a given background image.
We observe that competitive retrieval performance could be achieved by using a discriminator to predict the compatibility of composite image.
We propose a novel FOS method via distilling composite feature (DiscoFOS)
arXiv Detail & Related papers (2023-08-09T14:43:10Z) - MegaPose: 6D Pose Estimation of Novel Objects via Render & Compare [84.80956484848505]
MegaPose is a method to estimate the 6D pose of novel objects, that is, objects unseen during training.
We present a 6D pose refiner based on a render&compare strategy which can be applied to novel objects.
Second, we introduce a novel approach for coarse pose estimation which leverages a network trained to classify whether the pose error between a synthetic rendering and an observed image of the same object can be corrected by the refiner.
arXiv Detail & Related papers (2022-12-13T19:30:03Z) - FewSOL: A Dataset for Few-Shot Object Learning in Robotic Environments [21.393674766169543]
We introduce the Few-Shot Object Learning dataset for object recognition with a few images per object.
We captured 336 real-world objects with 9 RGB-D images per object from different views.
The evaluation results show that there is still a large margin to be improved for few-shot object classification in robotic environments.
arXiv Detail & Related papers (2022-07-06T05:57:24Z) - Discovering Objects that Can Move [55.743225595012966]
We study the problem of object discovery -- separating objects from the background without manual labels.
Existing approaches utilize appearance cues, such as color, texture, and location, to group pixels into object-like regions.
We choose to focus on dynamic objects -- entities that can move independently in the world.
arXiv Detail & Related papers (2022-03-18T21:13:56Z) - Making Images Real Again: A Comprehensive Survey on Deep Image Composition [34.09380539557308]
Image composition task could be into multiple sub-tasks, in which each sub-task targets at one or more issues.
In this paper, we conduct comprehensive survey over the sub-tasks and blending of image composition.
For each one, we summarize the existing methods, available datasets, and common evaluation metrics.
arXiv Detail & Related papers (2021-06-28T09:09:14Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z) - A Simple and Effective Use of Object-Centric Images for Long-Tailed
Object Detection [56.82077636126353]
We take advantage of object-centric images to improve object detection in scene-centric images.
We present a simple yet surprisingly effective framework to do so.
Our approach can improve the object detection (and instance segmentation) accuracy of rare objects by 50% (and 33%) relatively.
arXiv Detail & Related papers (2021-02-17T17:27:21Z) - Adversarial Image Composition with Auxiliary Illumination [53.89445873577062]
We propose an Adversarial Image Composition Net (AIC-Net) that achieves realistic image composition.
A novel branched generation mechanism is proposed, which disentangles the generation of shadows and the transfer of foreground styles.
Experiments on pedestrian and car composition tasks show that the proposed AIC-Net achieves superior composition performance.
arXiv Detail & Related papers (2020-09-17T12:58:16Z) - Object-Centric Image Generation from Layouts [93.10217725729468]
We develop a layout-to-image-generation method to generate complex scenes with multiple objects.
Our method learns representations of the spatial relationships between objects in the scene, which lead to our model's improved layout-fidelity.
We introduce SceneFID, an object-centric adaptation of the popular Fr'echet Inception Distance metric, that is better suited for multi-object images.
arXiv Detail & Related papers (2020-03-16T21:40:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.