Human-Aware Motion Deblurring
- URL: http://arxiv.org/abs/2001.06816v1
- Date: Sun, 19 Jan 2020 12:16:39 GMT
- Title: Human-Aware Motion Deblurring
- Authors: Ziyi Shen, Wenguan Wang, Xiankai Lu, Jianbing Shen, Haibin Ling,
Tingfa Xu, and Ling Shao
- Abstract summary: This paper proposes a human-aware deblurring model that disentangles the motion blur between foreground (FG) humans and background (BG)
The proposed model is based on a triple-branch encoder-decoder architecture.
The proposed model is further endowed with a supervised, human-aware attention mechanism in an end-to-end fashion.
- Score: 197.53076361425363
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper proposes a human-aware deblurring model that disentangles the
motion blur between foreground (FG) humans and background (BG). The proposed
model is based on a triple-branch encoder-decoder architecture. The first two
branches are learned for sharpening FG humans and BG details, respectively;
while the third one produces global, harmonious results by comprehensively
fusing multi-scale deblurring information from the two domains. The proposed
model is further endowed with a supervised, human-aware attention mechanism in
an end-to-end fashion. It learns a soft mask that encodes FG human information
and explicitly drives the FG/BG decoder-branches to focus on their specific
domains. To further benefit the research towards Human-aware Image Deblurring,
we introduce a large-scale dataset, named HIDE, which consists of 8,422 blurry
and sharp image pairs with 65,784 densely annotated FG human bounding boxes.
HIDE is specifically built to span a broad range of scenes, human object sizes,
motion patterns, and background complexities. Extensive experiments on public
benchmarks and our dataset demonstrate that our model performs favorably
against the state-of-the-art motion deblurring methods, especially in capturing
semantic details.
Related papers
- HumanSAM: Classifying Human-centric Forgery Videos in Human Spatial, Appearance, and Motion Anomaly [15.347208661111198]
HumanSAM aims to classify humancentric forgeries into three distinct types of artifacts commonly observed in generated content.<n>HumanSAM yields promising results in comparison with state-of-the-art methods, both in binary and multi-class forgery classification.
arXiv Detail & Related papers (2025-07-26T12:03:47Z) - DCHM: Depth-Consistent Human Modeling for Multiview Detection [13.81157120485514]
Multiview pedestrian detection typically involves two stages: human modeling and pedestrian localization.<n>We propose Depth-Consistent Human Modeling (DCHM), a framework designed for consistent depth estimation and multiview fusion.<n>Our proposed pipeline with superpixel-wise Gaussian Splatting achieves multiview depth consistency in sparse-view, large-scaled, and crowded scenarios.
arXiv Detail & Related papers (2025-07-19T06:37:14Z) - FAMOUS: High-Fidelity Monocular 3D Human Digitization Using View Synthesis [51.193297565630886]
The challenge of accurately inferring texture remains, particularly in obscured areas such as the back of a person in frontal-view images.
This limitation in texture prediction largely stems from the scarcity of large-scale and diverse 3D datasets.
We propose leveraging extensive 2D fashion datasets to enhance both texture and shape prediction in 3D human digitization.
arXiv Detail & Related papers (2024-10-13T01:25:05Z) - HOIMotion: Forecasting Human Motion During Human-Object Interactions Using Egocentric 3D Object Bounding Boxes [10.237077867790612]
We present HOIMotion, a novel approach for human motion forecasting during human-object interactions.
Our method integrates information about past body poses and egocentric 3D object bounding boxes.
We show that HOIMotion consistently outperforms state-of-the-art methods by a large margin.
arXiv Detail & Related papers (2024-07-02T19:58:35Z) - HyperHuman: Hyper-Realistic Human Generation with Latent Structural Diffusion [114.15397904945185]
We propose a unified framework, HyperHuman, that generates in-the-wild human images of high realism and diverse layouts.
Our model enforces the joint learning of image appearance, spatial relationship, and geometry in a unified network.
Our framework yields the state-of-the-art performance, generating hyper-realistic human images under diverse scenarios.
arXiv Detail & Related papers (2023-10-12T17:59:34Z) - Double-chain Constraints for 3D Human Pose Estimation in Images and
Videos [21.42410292863492]
Reconstructing 3D poses from 2D poses lacking depth information is challenging due to the complexity and diversity of human motion.
We propose a novel model, called Double-chain Graph Convolutional Transformer (DC-GCT), to constrain the pose.
We show that DC-GCT achieves state-of-the-art performance on two challenging datasets.
arXiv Detail & Related papers (2023-08-10T02:41:18Z) - Towards Hard-pose Virtual Try-on via 3D-aware Global Correspondence
Learning [70.75369367311897]
3D-aware global correspondences are reliable flows that jointly encode global semantic correlations, local deformations, and geometric priors of 3D human bodies.
An adversarial generator takes the garment warped by the 3D-aware flow, and the image of the target person as inputs, to synthesize the photo-realistic try-on result.
arXiv Detail & Related papers (2022-11-25T12:16:21Z) - MUG: Multi-human Graph Network for 3D Mesh Reconstruction from 2D Pose [20.099670445427964]
Reconstructing multi-human body mesh from a single monocular image is an important but challenging computer vision problem.
In this work, through a single graph neural network, we construct coherent multi-human meshes using only multi-human 2D pose as input.
arXiv Detail & Related papers (2022-05-25T08:54:52Z) - Image Comes Dancing with Collaborative Parsing-Flow Video Synthesis [124.48519390371636]
Transfering human motion from a source to a target person poses great potential in computer vision and graphics applications.
Previous work has either relied on crafted 3D human models or trained a separate model specifically for each target person.
This work studies a more general setting, in which we aim to learn a single model to parsimoniously transfer motion from a source video to any target person.
arXiv Detail & Related papers (2021-10-27T03:42:41Z) - Appearance Consensus Driven Self-Supervised Human Mesh Recovery [67.20942777949793]
We present a self-supervised human mesh recovery framework to infer human pose and shape from monocular images.
We achieve state-of-the-art results on the standard model-based 3D pose estimation benchmarks.
The resulting colored mesh prediction opens up the usage of our framework for a variety of appearance-related tasks beyond the pose and shape estimation.
arXiv Detail & Related papers (2020-08-04T05:40:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.