File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1109/CVPR42600.2020.00708
- Scopus: eid_2-s2.0-85094681073
- WOS: WOS:000620679507033
- Find via
Supplementary
- Citations:
- Appears in Collections:
Conference Paper: 3D Human Mesh Regression with Dense Correspondence
Title | 3D Human Mesh Regression with Dense Correspondence |
---|---|
Authors | |
Keywords | Three-dimensional displays Image reconstruction Solid modeling Biological system modeling Surface reconstruction |
Issue Date | 2020 |
Publisher | IEEE Computer Society. The Journal's web site is located at http://ieeexplore.ieee.org/xpl/conhome.jsp?punumber=1000147 |
Citation | Proceedings of IEEE/CVF International Conference on Computer Vision and Pattern Recognition (CVPR 2020), Seattle, USA, 13-19 June 2020, p. 7052-7061 How to Cite? |
Abstract | Estimating 3D mesh of the human body from a single 2D image is an important task with many applications such as augmented reality and Human-Robot interaction. However, prior works reconstructed 3D mesh from global image feature extracted by using convolutional neural network
(CNN), where the dense correspondences between the mesh surface and the image pixels are missing, leading to suboptimal solution. This paper proposes a model-free 3D human mesh estimation framework, named DecoMR, which explicitly establishes the dense correspondence between the
mesh and the local image features in the UV space (i.e. a 2D space used for texture mapping of 3D mesh). DecoMR first predicts pixel-to-surface dense correspondence map (i.e., IUV image), with which we transfer local features from the image space to the UV space. Then the transferred local image features are processed in the UV space to regress a location map, which is well aligned with transferred features. Finally we reconstruct 3D human mesh from the regressed location map with a predefined mapping function. We also observe that the existing discontinuous UV map are unfriendly to the learning
of network. Therefore, we propose a novel UV map that maintains most of the neighboring relations on the original mesh surface. Experiments demonstrate that our proposed local feature alignment and continuous UV map outperforms existing 3D mesh based methods on multiple public benchmarks. Code will be made available at https://github.com/zengwang430521/DecoMR. |
Description | Session: Poster 2.2 — Face, Gesture, and Body Pose; Motion and Tracking; Representation Learning - Poster no. 94; Paper ID 6333 CVPR 2020 held virtually due to COVID-19 |
Persistent Identifier | http://hdl.handle.net/10722/284160 |
ISSN | 2023 SCImago Journal Rankings: 10.331 |
ISI Accession Number ID |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Zeng, W | - |
dc.contributor.author | Ouyang, W | - |
dc.contributor.author | Luo, P | - |
dc.contributor.author | Liu, W | - |
dc.contributor.author | Wang, X | - |
dc.date.accessioned | 2020-07-20T05:56:33Z | - |
dc.date.available | 2020-07-20T05:56:33Z | - |
dc.date.issued | 2020 | - |
dc.identifier.citation | Proceedings of IEEE/CVF International Conference on Computer Vision and Pattern Recognition (CVPR 2020), Seattle, USA, 13-19 June 2020, p. 7052-7061 | - |
dc.identifier.issn | 1063-6919 | - |
dc.identifier.uri | http://hdl.handle.net/10722/284160 | - |
dc.description | Session: Poster 2.2 — Face, Gesture, and Body Pose; Motion and Tracking; Representation Learning - Poster no. 94; Paper ID 6333 | - |
dc.description | CVPR 2020 held virtually due to COVID-19 | - |
dc.description.abstract | Estimating 3D mesh of the human body from a single 2D image is an important task with many applications such as augmented reality and Human-Robot interaction. However, prior works reconstructed 3D mesh from global image feature extracted by using convolutional neural network (CNN), where the dense correspondences between the mesh surface and the image pixels are missing, leading to suboptimal solution. This paper proposes a model-free 3D human mesh estimation framework, named DecoMR, which explicitly establishes the dense correspondence between the mesh and the local image features in the UV space (i.e. a 2D space used for texture mapping of 3D mesh). DecoMR first predicts pixel-to-surface dense correspondence map (i.e., IUV image), with which we transfer local features from the image space to the UV space. Then the transferred local image features are processed in the UV space to regress a location map, which is well aligned with transferred features. Finally we reconstruct 3D human mesh from the regressed location map with a predefined mapping function. We also observe that the existing discontinuous UV map are unfriendly to the learning of network. Therefore, we propose a novel UV map that maintains most of the neighboring relations on the original mesh surface. Experiments demonstrate that our proposed local feature alignment and continuous UV map outperforms existing 3D mesh based methods on multiple public benchmarks. Code will be made available at https://github.com/zengwang430521/DecoMR. | - |
dc.language | eng | - |
dc.publisher | IEEE Computer Society. The Journal's web site is located at http://ieeexplore.ieee.org/xpl/conhome.jsp?punumber=1000147 | - |
dc.relation.ispartof | IEEE Conference on Computer Vision and Pattern Recognition. Proceedings | - |
dc.rights | IEEE Conference on Computer Vision and Pattern Recognition. Proceedings. Copyright © IEEE Computer Society. | - |
dc.rights | ©2020 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. | - |
dc.subject | Three-dimensional displays | - |
dc.subject | Image reconstruction | - |
dc.subject | Solid modeling | - |
dc.subject | Biological system modeling | - |
dc.subject | Surface reconstruction | - |
dc.title | 3D Human Mesh Regression with Dense Correspondence | - |
dc.type | Conference_Paper | - |
dc.identifier.email | Luo, P: pluo@hku.hk | - |
dc.identifier.authority | Luo, P=rp02575 | - |
dc.description.nature | link_to_subscribed_fulltext | - |
dc.identifier.doi | 10.1109/CVPR42600.2020.00708 | - |
dc.identifier.scopus | eid_2-s2.0-85094681073 | - |
dc.identifier.hkuros | 311020 | - |
dc.identifier.spage | 7052 | - |
dc.identifier.epage | 7061 | - |
dc.identifier.isi | WOS:000620679507033 | - |
dc.publisher.place | United States | - |
dc.identifier.issnl | 1063-6919 | - |