STG-Avatar: Animatable Human Avatars via Spacetime Gaussian
- URL: http://arxiv.org/abs/2510.22140v1
- Date: Sat, 25 Oct 2025 03:23:38 GMT
- Title: STG-Avatar: Animatable Human Avatars via Spacetime Gaussian
- Authors: Guangan Jiang, Tianzi Zhang, Dong Li, Zhenjun Zhao, Haoang Li, Mingrui Li, Hongyu Wang,
- Abstract summary: We present STG-Avatar, a 3DGS-based framework for high-fidelity animatable human avatar reconstruction.<n>LBS enables real-time skeletal control by driving global pose transformations.<n>Our method consistently outperforms state-of-the-art baselines in both reconstruction quality and operational efficiency.
- Score: 14.962899842675304
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Realistic animatable human avatars from monocular videos are crucial for advancing human-robot interaction and enhancing immersive virtual experiences. While recent research on 3DGS-based human avatars has made progress, it still struggles with accurately representing detailed features of non-rigid objects (e.g., clothing deformations) and dynamic regions (e.g., rapidly moving limbs). To address these challenges, we present STG-Avatar, a 3DGS-based framework for high-fidelity animatable human avatar reconstruction. Specifically, our framework introduces a rigid-nonrigid coupled deformation framework that synergistically integrates Spacetime Gaussians (STG) with linear blend skinning (LBS). In this hybrid design, LBS enables real-time skeletal control by driving global pose transformations, while STG complements it through spacetime adaptive optimization of 3D Gaussians. Furthermore, we employ optical flow to identify high-dynamic regions and guide the adaptive densification of 3D Gaussians in these regions. Experimental results demonstrate that our method consistently outperforms state-of-the-art baselines in both reconstruction quality and operational efficiency, achieving superior quantitative metrics while retaining real-time rendering capabilities. Our code is available at https://github.com/jiangguangan/STG-Avatar
Related papers
- CAG-Avatar: Cross-Attention Guided Gaussian Avatars for High-Fidelity Head Reconstruction [7.698661374784336]
Animation techniques often rely on a "one-size-fits-all" global tuning approach.<n>We introduce Conditionally- Adaptive Fusion Module built on cross-attention.<n>Experiments confirm a significant improvement in reconstruction fidelity, particularly for challenging regions such as teeth.
arXiv Detail & Related papers (2026-01-21T10:22:53Z) - HoliGS: Holistic Gaussian Splatting for Embodied View Synthesis [59.25751939710903]
We propose a novel deformable Gaussian splatting framework that addresses embodied view synthesis from long monocular RGB videos.<n>Our method leverages invertible Gaussian Splatting deformation networks to reconstruct large-scale, dynamic environments accurately.<n>Results highlight a practical and scalable solution for EVS in real-world scenarios.
arXiv Detail & Related papers (2025-06-24T03:54:40Z) - Parametric Gaussian Human Model: Generalizable Prior for Efficient and Realistic Human Avatar Modeling [32.480049588166544]
Photo and animatable human avatars are a key enabler for virtual/augmented reality, telepresence, and digital entertainment.<n>We present the Parametric Gaussian Human Model (PGHM), a generalizable and efficient framework that integrates human priors into 3DGS.<n>Experiments show that PGHM is significantly more efficient than optimization-from-scratch methods, requiring only approximately 20 minutes per subject to produce avatars with comparable visual quality.
arXiv Detail & Related papers (2025-06-07T03:53:30Z) - RealityAvatar: Towards Realistic Loose Clothing Modeling in Animatable 3D Gaussian Avatars [4.332718737928592]
We propose RealityAvatar, an efficient framework for high-fidelity digital human modeling, specifically targeting loosely dressed avatars.<n>By incorporating a motion trend module and a latentbone encoder, we explicitly model pose-dependent deformations and temporal variations in clothing behavior.<n>Our method significantly enhances structural fidelity and perceptual quality in dynamic human reconstruction, particularly in non-rigid regions.
arXiv Detail & Related papers (2025-04-02T09:59:12Z) - EVolSplat: Efficient Volume-based Gaussian Splatting for Urban View Synthesis [61.1662426227688]
Existing NeRF and 3DGS-based methods show promising results in achieving photorealistic renderings but require slow, per-scene optimization.<n>We introduce EVolSplat, an efficient 3D Gaussian Splatting model for urban scenes that works in a feed-forward manner.
arXiv Detail & Related papers (2025-03-26T02:47:27Z) - 2DGS-Avatar: Animatable High-fidelity Clothed Avatar via 2D Gaussian Splatting [10.935483693282455]
We propose 2DGS-Avatar, a novel approach for modeling animatable clothed avatars with high-fidelity and fast training performance.<n>Our method generates an avatar that can be driven by poses and rendered in real-time.<n>Compared to 3DGS-based methods, our 2DGS-Avatar retains the advantages of fast training and rendering while also capturing detailed, dynamic, and photo-realistic appearances.
arXiv Detail & Related papers (2025-03-04T09:57:24Z) - Sequential Gaussian Avatars with Hierarchical Motion Context [7.6736633105043515]
SMPL-driven 3DGS human avatars struggle to capture fine appearance details due to complex mapping from pose to appearance during fitting.<n>We propose SeqAvatar, which excavates the explicit 3DGS representation to better model human avatars based on a hierarchical motion context.<n>Our method significantly outperforms 3DGS-based approaches and renders human avatars rendering orders of magnitude faster than the latest NeRF-based models.
arXiv Detail & Related papers (2024-11-25T04:05:19Z) - Topology-aware Human Avatars with Semantically-guided Gaussian Splatting [18.421585526595944]
We propose SG-GS, which uses semantics-embedded 3D Gaussians, skeleton-driven rigid deformation, and non-rigid cloth dynamics deformation to create photo-realistic human avatars.
We employ a 3D network that integrates both topological and geometric associations for human avatar deformation.
arXiv Detail & Related papers (2024-08-19T02:58:20Z) - Spec-Gaussian: Anisotropic View-Dependent Appearance for 3D Gaussian Splatting [55.71424195454963]
Spec-Gaussian is an approach that utilizes an anisotropic spherical Gaussian appearance field instead of spherical harmonics.
Our experimental results demonstrate that our method surpasses existing approaches in terms of rendering quality.
This improvement extends the applicability of 3D GS to handle intricate scenarios with specular and anisotropic surfaces.
arXiv Detail & Related papers (2024-02-24T17:22:15Z) - ASH: Animatable Gaussian Splats for Efficient and Photoreal Human Rendering [62.81677824868519]
We propose an animatable Gaussian splatting approach for photorealistic rendering of dynamic humans in real-time.
We parameterize the clothed human as animatable 3D Gaussians, which can be efficiently splatted into image space to generate the final rendering.
We benchmark ASH with competing methods on pose-controllable avatars, demonstrating that our method outperforms existing real-time methods by a large margin and shows comparable or even better results than offline methods.
arXiv Detail & Related papers (2023-12-10T17:07:37Z) - GaussianAvatar: Towards Realistic Human Avatar Modeling from a Single Video via Animatable 3D Gaussians [51.46168990249278]
We present an efficient approach to creating realistic human avatars with dynamic 3D appearances from a single video.
GustafAvatar is validated on both the public dataset and our collected dataset.
arXiv Detail & Related papers (2023-12-04T18:55:45Z) - SC-GS: Sparse-Controlled Gaussian Splatting for Editable Dynamic Scenes [59.23385953161328]
Novel view synthesis for dynamic scenes is still a challenging problem in computer vision and graphics.
We propose a new representation that explicitly decomposes the motion and appearance of dynamic scenes into sparse control points and dense Gaussians.
Our method can enable user-controlled motion editing while retaining high-fidelity appearances.
arXiv Detail & Related papers (2023-12-04T11:57:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.