E3Gen: Efficient, Expressive and Editable Avatars Generation

Weitian Zhang1
Yunhui Liu2
Xingdong Sheng2
1Shanghai Jiao Tong University 2Lenovo Research

Our novel method, E3Gen, demonstrates its capability of generating high-fidelity avatars with detailed cloth wrinkles and achieves a rendering resolution of 10242 in real time. Furthermore, our method offers comprehensive control over camera views, and full-body poses, and supports attribute transfer and local editing.


This paper aims to introduce 3D Gaussian for efficient, expressive, and editable digital avatar generation. This task faces two major challenges: (1) The unstructured nature of 3D Gaussian makes it incompatible with current generation pipelines; (2) the expressive animation of 3D Gaussian in a generative setting that involves training with multiple subjects remains unexplored. In this paper, we propose a novel avatar generation method named E3Gen, to effectively address these challenges. First, we propose a novel generative UV features plane representation that encodes unstructured 3D Gaussian onto a structured 2D UV space defined by the SMPL-X parametric model. This novel representation not only preserves the representation ability of the original 3D Gaussian but also introduces a shared structure among subjects to enable generative learning of the diffusion model. To tackle the second challenge, we propose a part-aware deformation module to achieve robust and accurate full-body expressive pose control. Extensive experiments demonstrate that our method achieves superior performance in avatar generation and enables expressive full-body pose control and editing.

Full Video

Method Overview

Our approach utilizes a single-stage diffusion model to simultaneously train the denoising and fitting process. The UV features plane, xi, is randomly initialized and optimized by both processes. In the denoising process, noise is added to the UV features plane and then denoised following a v-parameterization scheme using a denoising UNet. In the fitting process, the UV features plane is decoded into Gaussian Attribute maps, which are used to generate a 3D-Gaussian-based avatar in canonical space by fetching the corresponding attributes for the initialized Gaussian primitive. Finally, a part-aware deformation module is employed to deform the avatar into targeted pose based on SMPL-X parameters.

Random Generation

Our method can generate digital avatars with complex appearance and rich details, clearly exhibiting fingers and clothing wrinkles.

Novel Pose Animation

Our method exhibit robustly animation results to these challenging novel poses.


Our method enables local editing and partial attribute transfer. The edited results also support pose control.


    title={$E^{3}$Gen: Efficient, Expressive and Editable Avatars Generation}, 
    author={Weitian Zhang and Yichao Yan and Yunhui Liu and Xingdong Sheng and Xiaokang Yang},
    journal={arXiv preprint arXiv:2405.19203},