ILSH: The Imperial Light-Stage Head Dataset for Human Head View
Synthesis
- URL: http://arxiv.org/abs/2310.03952v1
- Date: Fri, 6 Oct 2023 00:32:36 GMT
- Title: ILSH: The Imperial Light-Stage Head Dataset for Human Head View
Synthesis
- Authors: Jiali Zheng, Youngkyoon Jang, Athanasios Papaioannou, Christos
Kampouris, Rolandos Alexandros Potamias, Foivos Paraperas Papantoniou,
Efstathios Galanakis, Ales Leonardis, Stefanos Zafeiriou
- Abstract summary: Imperial Light-Stage Head dataset is a novel dataset designed to support view synthesis academic challenges for human heads.
This paper details the setup of a light-stage specifically designed to capture high-resolution (4K) human head images.
In addition to the data collection, we address the split of the dataset into train, validation, and test sets.
- Score: 42.81410101705251
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper introduces the Imperial Light-Stage Head (ILSH) dataset, a novel
light-stage-captured human head dataset designed to support view synthesis
academic challenges for human heads. The ILSH dataset is intended to facilitate
diverse approaches, such as scene-specific or generic neural rendering,
multiple-view geometry, 3D vision, and computer graphics, to further advance
the development of photo-realistic human avatars. This paper details the setup
of a light-stage specifically designed to capture high-resolution (4K) human
head images and describes the process of addressing challenges (preprocessing,
ethical issues) in collecting high-quality data. In addition to the data
collection, we address the split of the dataset into train, validation, and
test sets. Our goal is to design and support a fair view synthesis challenge
task for this novel dataset, such that a similar level of performance can be
maintained and expected when using the test set, as when using the validation
set. The ILSH dataset consists of 52 subjects captured using 24 cameras with
all 82 lighting sources turned on, resulting in a total of 1,248 close-up head
images, border masks, and camera pose pairs.
Related papers
- VGGHeads: A Large-Scale Synthetic Dataset for 3D Human Heads [18.62716110331954]
VGGHeads is a large scale synthetic dataset generated with diffusion models for human head detection and 3D mesh estimation.
Our dataset comprises over 1 million high-resolution images, each annotated with detailed 3D head meshes, facial landmarks, and bounding boxes.
arXiv Detail & Related papers (2024-07-25T17:58:17Z) - 360 in the Wild: Dataset for Depth Prediction and View Synthesis [66.58513725342125]
We introduce a large scale 360$circ$ videos dataset in the wild.
This dataset has been carefully scraped from the Internet and has been captured from various locations worldwide.
Each of the 25K images constituting our dataset is provided with its respective camera's pose and depth map.
arXiv Detail & Related papers (2024-06-27T05:26:38Z) - LWIRPOSE: A novel LWIR Thermal Image Dataset and Benchmark [9.679771580702258]
This dataset comprises over 2,400 high-quality LWIR (thermal) images.
Each image is meticulously annotated with 2D human poses, offering a valuable resource for researchers and practitioners.
We benchmark state-of-the-art pose estimation methods on the dataset to showcase its potential.
arXiv Detail & Related papers (2024-04-16T01:49:35Z) - Headset: Human emotion awareness under partial occlusions multimodal
dataset [19.57427512904342]
We present a new multimodal database to help advance the development of immersive technologies.
Our proposed database provides ethically compliant and diverse volumetric data, in particular 27 participants displaying posed facial expressions and subtle body movements while speaking, plus 11 participants wearing head-mounted displays (HMDs)
The dataset can be helpful in the evaluation and performance testing of various XR algorithms, including but not limited to facial expression recognition and reconstruction, facial reenactment, and volumetric video.
arXiv Detail & Related papers (2024-02-14T11:42:15Z) - DNA-Rendering: A Diverse Neural Actor Repository for High-Fidelity
Human-centric Rendering [126.00165445599764]
We present DNA-Rendering, a large-scale, high-fidelity repository of human performance data for neural actor rendering.
Our dataset contains over 1500 human subjects, 5000 motion sequences, and 67.5M frames' data volume.
We construct a professional multi-view system to capture data, which contains 60 synchronous cameras with max 4096 x 3000 resolution, 15 fps speed, and stern camera calibration steps.
arXiv Detail & Related papers (2023-07-19T17:58:03Z) - RenderMe-360: A Large Digital Asset Library and Benchmarks Towards
High-fidelity Head Avatars [157.82758221794452]
We present RenderMe-360, a comprehensive 4D human head dataset to drive advance in head avatar research.
It contains massive data assets, with 243+ million complete head frames, and over 800k video sequences from 500 different identities.
Based on the dataset, we build a comprehensive benchmark for head avatar research, with 16 state-of-the-art methods performed on five main tasks.
arXiv Detail & Related papers (2023-05-22T17:54:01Z) - StyleGAN-Human: A Data-Centric Odyssey of Human Generation [96.7080874757475]
This work takes a data-centric perspective and investigates multiple critical aspects in "data engineering"
We collect and annotate a large-scale human image dataset with over 230K samples capturing diverse poses and textures.
We rigorously investigate three essential factors in data engineering for StyleGAN-based human generation, namely data size, data distribution, and data alignment.
arXiv Detail & Related papers (2022-04-25T17:55:08Z) - UET-Headpose: A sensor-based top-view head pose dataset [0.0]
We introduce a new approach with efficient cost and easy setup to collecting head pose images.
This method uses an absolute orientation sensor instead of Depth cameras to be set up quickly and small cost.
We also introduce the full-range model called FSANet-Wide, which significantly outperforms head pose estimation results by the UET-Headpose dataset.
arXiv Detail & Related papers (2021-11-13T04:54:20Z) - Hypersim: A Photorealistic Synthetic Dataset for Holistic Indoor Scene
Understanding [8.720130442653575]
Hypersim is a synthetic dataset for holistic indoor scene understanding.
We generate 77,400 images of 461 indoor scenes with detailed per-pixel labels and corresponding ground truth geometry.
arXiv Detail & Related papers (2020-11-04T20:12:07Z) - SelfPose: 3D Egocentric Pose Estimation from a Headset Mounted Camera [97.0162841635425]
We present a solution to egocentric 3D body pose estimation from monocular images captured from downward looking fish-eye cameras installed on the rim of a head mounted VR device.
This unusual viewpoint leads to images with unique visual appearance, with severe self-occlusions and perspective distortions.
We propose an encoder-decoder architecture with a novel multi-branch decoder designed to account for the varying uncertainty in 2D predictions.
arXiv Detail & Related papers (2020-11-02T16:18:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.