File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: DreamComposer: Controllable 3D Object Generation via Multi-View Conditions

TitleDreamComposer: Controllable 3D Object Generation via Multi-View Conditions
Authors
Issue Date17-Jun-2024
Abstract

Utilizing pretrained 2D large-scale generative models, recent works are capable of generating high-quality novel views from a single in-the-wild image. However, due to the lack of information from multiple views, these works encounter difficulties in generating controllable novel views. In this paper, we present DreamComposer, a flexible and scalable framework that can enhance existing view-aware diffusion models by injecting multi-view conditions. Specifically, DreamComposer first uses a view-aware 3D lifting module to obtain 3D representations of an object from multiple views. Then, it renders the latent features of the target view from 3D representations with the multi-view feature fusion module. Finally the target view features extracted from multi-view inputs are injected into a pretrained diffusion model. Experiments show that DreamComposer is compatible with state-of-the-art diffusion models for zero-shot novel view synthesis, further enhancing them to generate high-fidelity novel view images with multi-view conditions, ready for controllable 3D object reconstruction and various other applications.


Persistent Identifierhttp://hdl.handle.net/10722/350520

 

DC FieldValueLanguage
dc.contributor.authorYang, Yunhan-
dc.contributor.authorHuang, Yukun-
dc.contributor.authorWu, Xiaoyang-
dc.contributor.authorGuo, Yuanchen-
dc.contributor.authorZhang, Song-Hai-
dc.contributor.authorZhao, Hengshuang-
dc.contributor.authorHe, Tong-
dc.contributor.authorLiu, Xihui-
dc.date.accessioned2024-10-29T00:32:02Z-
dc.date.available2024-10-29T00:32:02Z-
dc.date.issued2024-06-17-
dc.identifier.urihttp://hdl.handle.net/10722/350520-
dc.description.abstract<p>Utilizing pretrained 2D large-scale generative models, recent works are capable of generating high-quality novel views from a single in-the-wild image. However, due to the lack of information from multiple views, these works encounter difficulties in generating controllable novel views. In this paper, we present DreamComposer, a flexible and scalable framework that can enhance existing view-aware diffusion models by injecting multi-view conditions. Specifically, DreamComposer first uses a view-aware 3D lifting module to obtain 3D representations of an object from multiple views. Then, it renders the latent features of the target view from 3D representations with the multi-view feature fusion module. Finally the target view features extracted from multi-view inputs are injected into a pretrained diffusion model. Experiments show that DreamComposer is compatible with state-of-the-art diffusion models for zero-shot novel view synthesis, further enhancing them to generate high-fidelity novel view images with multi-view conditions, ready for controllable 3D object reconstruction and various other applications.</p>-
dc.languageeng-
dc.relation.ispartof2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (17/06/2024-21/06/2024, Seattle)-
dc.titleDreamComposer: Controllable 3D Object Generation via Multi-View Conditions-
dc.typeConference_Paper-
dc.identifier.doi10.1109/CVPR52733.2024.00775-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats