We reconstruct and animate inputs from the SHHQ dataset.
Although we do not aim to train a generative model, we observe some emerging capabilities. The avatar latent code can be edited and interpolated.
The video interpolates between training subjects.
We replace the face and shoes of the subject on the left with items from the identity on the right.0
Compared with IDOL (the state-of-the-art), DLA achieves higher-quality and more photorealistic avatars.
We also compare with DreamGaussian, SiTH, and SIFU in the paper.
We thank Shalini De Mello and Jenny Schmalfuss for their valuable inputs, and Jan Kautz for hosting Marcel's internship.
@misc{buehler2025dla,
title={Dream, Lift, Animate: From Single Images to Animatable Gaussian Avatars},
author={Marcel C. Buehler, Ye Yuan, Xueting Li, Yangyi Huang, Koki Nagano, Umar Iqbal},
year={2025},
eprint={XXXX.XXXXX},
archivePrefix={arXiv},
primaryClass={cs.CV}
}