File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: MVPortrait: Text-Guided Motion and Emotion Control for Multi-view Vivid Portrait Animation

TitleMVPortrait: Text-Guided Motion and Emotion Control for Multi-view Vivid Portrait Animation
Authors
Issue Date11-Jun-2025
Abstract

Recent portrait animation methods have made significant strides in generating realistic lip synchronization. However, they often lack explicit control over head movements and facial expressions, and cannot produce videos from multiple viewpoints, resulting in less controllable and expressive animations. Moreover, text-guided portrait animation remains underexplored, despite its user-friendly nature. We present a novel two-stage text-guided framework, MVPortrait (Multi-view Vivid Portrait), to generate expressive multi-view portrait animations that faithfully capture the described motion and emotion. MVPortrait is the first to introduce FLAME as an intermediate representation, effectively embedding facial movements, expressions, and view transformations within its parameter space. In the first stage, we separately train the FLAME motion and emotion diffusion models based on text input. In the second stage, we train a multi-view video generation model conditioned on a reference portrait image and multi-view FLAME rendering sequences from the first stage. Experimental results exhibit that MVPortrait outperforms existing methods in terms of motion and emotion control, as well as view consistency. Furthermore, by leveraging FLAME as a bridge, MVPortrait becomes the first controllable portrait animation framework that is compatible with text, speech, and video as driving signals.


Persistent Identifierhttp://hdl.handle.net/10722/358646

 

DC FieldValueLanguage
dc.contributor.authorLin, Yukang-
dc.contributor.authorFung, Hokit-
dc.contributor.authorXu, Jianjin-
dc.contributor.authorRen, Zeping-
dc.contributor.authorLau, Adela Sau Mui-
dc.contributor.authorYin, Guosheng-
dc.contributor.authorLi, Xiu-
dc.date.accessioned2025-08-13T07:47:11Z-
dc.date.available2025-08-13T07:47:11Z-
dc.date.issued2025-06-11-
dc.identifier.urihttp://hdl.handle.net/10722/358646-
dc.description.abstract<p>Recent portrait animation methods have made significant strides in generating realistic lip synchronization. However, they often lack explicit control over head movements and facial expressions, and cannot produce videos from multiple viewpoints, resulting in less controllable and expressive animations. Moreover, text-guided portrait animation remains underexplored, despite its user-friendly nature. We present a novel two-stage text-guided framework, MVPortrait (Multi-view Vivid Portrait), to generate expressive multi-view portrait animations that faithfully capture the described motion and emotion. MVPortrait is the first to introduce FLAME as an intermediate representation, effectively embedding facial movements, expressions, and view transformations within its parameter space. In the first stage, we separately train the FLAME motion and emotion diffusion models based on text input. In the second stage, we train a multi-view video generation model conditioned on a reference portrait image and multi-view FLAME rendering sequences from the first stage. Experimental results exhibit that MVPortrait outperforms existing methods in terms of motion and emotion control, as well as view consistency. Furthermore, by leveraging FLAME as a bridge, MVPortrait becomes the first controllable portrait animation framework that is compatible with text, speech, and video as driving signals.<br></p>-
dc.languageeng-
dc.relation.ispartofIEEE Conference on Computer Vision and Pattern Recognition 2025 (11/06/2025-15/06/2025, Nashville TN)-
dc.titleMVPortrait: Text-Guided Motion and Emotion Control for Multi-view Vivid Portrait Animation-
dc.typeConference_Paper-
dc.identifier.doi10.48550/arXiv.2503.19383-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats