ChaLearn Looking at People: Inpainting and Denoising challenges
- URL: http://arxiv.org/abs/2106.13071v1
- Date: Thu, 24 Jun 2021 14:57:21 GMT
- Title: ChaLearn Looking at People: Inpainting and Denoising challenges
- Authors: Sergio Escalera and Marti Soler and Stephane Ayache and Umut Guclu and
Jun Wan and Meysam Madadi and Xavier Baro and Hugo Jair Escalante and
Isabelle Guyon
- Abstract summary: This chapter describes the design of an academic competition focusing on inpainting of images and video sequences.
The ChaLearn Looking at People Inpainting Challenge aimed at advancing the state of the art on visual inpainting.
Three tracks were proposed in which visual inpainting might be helpful but still challenging: human body pose estimation, text overlays removal and fingerprint denoising.
- Score: 41.481257371694284
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dealing with incomplete information is a well studied problem in the context
of machine learning and computational intelligence. However, in the context of
computer vision, the problem has only been studied in specific scenarios (e.g.,
certain types of occlusions in specific types of images), although it is common
to have incomplete information in visual data. This chapter describes the
design of an academic competition focusing on inpainting of images and video
sequences that was part of the competition program of WCCI2018 and had a
satellite event collocated with ECCV2018. The ChaLearn Looking at People
Inpainting Challenge aimed at advancing the state of the art on visual
inpainting by promoting the development of methods for recovering missing and
occluded information from images and video. Three tracks were proposed in which
visual inpainting might be helpful but still challenging: human body pose
estimation, text overlays removal and fingerprint denoising. This chapter
describes the design of the challenge, which includes the release of three
novel datasets, and the description of evaluation metrics, baselines and
evaluation protocol. The results of the challenge are analyzed and discussed in
detail and conclusions derived from this event are outlined.
Related papers
- Deep Learning-based Image and Video Inpainting: A Survey [47.53641171826598]
This paper comprehensively reviews the deep learning-based methods for image and video inpainting.
We sort existing methods into different categories from the perspective of their high-level inpainting pipeline.
We present evaluation metrics for low-level pixel and high-level perceptional similarity, conduct a performance evaluation, and discuss the strengths and weaknesses of representative inpainting methods.
arXiv Detail & Related papers (2024-01-07T05:50:12Z) - NICE: CVPR 2023 Challenge on Zero-shot Image Captioning [149.28330263581012]
NICE project is designed to challenge the computer vision community to develop robust image captioning models.
Report includes information on the newly proposed NICE dataset, evaluation methods, challenge results, and technical details of top-ranking entries.
arXiv Detail & Related papers (2023-09-05T05:32:19Z) - Free-ATM: Exploring Unsupervised Learning on Diffusion-Generated Images
with Free Attention Masks [64.67735676127208]
Text-to-image diffusion models have shown great potential for benefiting image recognition.
Although promising, there has been inadequate exploration dedicated to unsupervised learning on diffusion-generated images.
We introduce customized solutions by fully exploiting the aforementioned free attention masks.
arXiv Detail & Related papers (2023-08-13T10:07:46Z) - Exploring CLIP for Assessing the Look and Feel of Images [87.97623543523858]
We introduce Contrastive Language-Image Pre-training (CLIP) models for assessing both the quality perception (look) and abstract perception (feel) of images in a zero-shot manner.
Our results show that CLIP captures meaningful priors that generalize well to different perceptual assessments.
arXiv Detail & Related papers (2022-07-25T17:58:16Z) - Describing image focused in cognitive and visual details for visually
impaired people: An approach to generating inclusive paragraphs [2.362412515574206]
There is a lack of services that support specific tasks, such as understanding the image context presented in online content, e.g., webinars.
We propose an approach for generating context of webinar images combining a dense captioning technique with a set of filters, to fit the captions in our domain, and a language model for the abstractive summary task.
arXiv Detail & Related papers (2022-02-10T21:20:53Z) - Deep Image Deblurring: A Survey [165.32391279761006]
Deblurring is a classic problem in low-level computer vision, which aims to recover a sharp image from a blurred input image.
Recent advances in deep learning have led to significant progress in solving this problem.
arXiv Detail & Related papers (2022-01-26T01:31:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.