Dream, Lift, Animate: From Single Images to Animatable Gaussian Avatars
Reading time: 2 minute
...
📝 Original Info
- Title: Dream, Lift, Animate: From Single Images to Animatable Gaussian Avatars
- ArXiv ID: 2507.15979
- Date: 2025-07-21
- Authors: ** 정보 없음 (제공된 텍스트에 저자 정보가 포함되어 있지 않습니다.) **
📝 Abstract
We introduce Dream, Lift, Animate (DLA), a novel framework that reconstructs animatable 3D human avatars from a single image. This is achieved by leveraging multi-view generation, 3D Gaussian lifting, and pose-aware UV-space mapping of 3D Gaussians. Given an image, we first dream plausible multi-views using a video diffusion model, capturing rich geometric and appearance details. These views are then lifted into unstructured 3D Gaussians. To enable animation, we propose a transformer-based encoder that models global spatial relationships and projects these Gaussians into a structured latent representation aligned with the UV space of a parametric body model. This latent code is decoded into UV-space Gaussians that can be animated via body-driven deformation and rendered conditioned on pose and viewpoint. By anchoring Gaussians to the UV manifold, our method ensures consistency during animation while preserving fine visual details. DLA enables real-time rendering and intuitive editing without requiring post-processing. Our method outperforms state-of-the-art approaches on the ActorsHQ and 4D-Dress datasets in both perceptual quality and photometric accuracy. By combining the generative strengths of video diffusion models with a pose-aware UV-space Gaussian mapping, DLA bridges the gap between unstructured 3D representations and high-fidelity, animation-ready avatars.💡 Deep Analysis
📄 Full Content
Reference
This content is AI-processed based on open access ArXiv data.