File Download
There are no files associated with this item.
Supplementary
-
Citations:
- Appears in Collections:
Conference Paper: DreamComposer: Controllable 3D Object Generation via Multi-View Conditions
Title | DreamComposer: Controllable 3D Object Generation via Multi-View Conditions |
---|---|
Authors | |
Issue Date | 17-Jun-2024 |
Abstract | Utilizing pretrained 2D large-scale generative models, recent works are capable of generating high-quality novel views from a single in-the-wild image. However, due to the lack of information from multiple views, these works encounter difficulties in generating controllable novel views. In this paper, we present DreamComposer, a flexible and scalable framework that can enhance existing view-aware diffusion models by injecting multi-view conditions. Specifically, DreamComposer first uses a view-aware 3D lifting module to obtain 3D representations of an object from multiple views. Then, it renders the latent features of the target view from 3D representations with the multi-view feature fusion module. Finally the target view features extracted from multi-view inputs are injected into a pretrained diffusion model. Experiments show that DreamComposer is compatible with state-of-the-art diffusion models for zero-shot novel view synthesis, further enhancing them to generate high-fidelity novel view images with multi-view conditions, ready for controllable 3D object reconstruction and various other applications. |
Persistent Identifier | http://hdl.handle.net/10722/350520 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Yang, Yunhan | - |
dc.contributor.author | Huang, Yukun | - |
dc.contributor.author | Wu, Xiaoyang | - |
dc.contributor.author | Guo, Yuanchen | - |
dc.contributor.author | Zhang, Song-Hai | - |
dc.contributor.author | Zhao, Hengshuang | - |
dc.contributor.author | He, Tong | - |
dc.contributor.author | Liu, Xihui | - |
dc.date.accessioned | 2024-10-29T00:32:02Z | - |
dc.date.available | 2024-10-29T00:32:02Z | - |
dc.date.issued | 2024-06-17 | - |
dc.identifier.uri | http://hdl.handle.net/10722/350520 | - |
dc.description.abstract | <p>Utilizing pretrained 2D large-scale generative models, recent works are capable of generating high-quality novel views from a single in-the-wild image. However, due to the lack of information from multiple views, these works encounter difficulties in generating controllable novel views. In this paper, we present DreamComposer, a flexible and scalable framework that can enhance existing view-aware diffusion models by injecting multi-view conditions. Specifically, DreamComposer first uses a view-aware 3D lifting module to obtain 3D representations of an object from multiple views. Then, it renders the latent features of the target view from 3D representations with the multi-view feature fusion module. Finally the target view features extracted from multi-view inputs are injected into a pretrained diffusion model. Experiments show that DreamComposer is compatible with state-of-the-art diffusion models for zero-shot novel view synthesis, further enhancing them to generate high-fidelity novel view images with multi-view conditions, ready for controllable 3D object reconstruction and various other applications.</p> | - |
dc.language | eng | - |
dc.relation.ispartof | 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (17/06/2024-21/06/2024, Seattle) | - |
dc.title | DreamComposer: Controllable 3D Object Generation via Multi-View Conditions | - |
dc.type | Conference_Paper | - |
dc.identifier.doi | 10.1109/CVPR52733.2024.00775 | - |