File Download
There are no files associated with this item.
Supplementary
-
Citations:
- Appears in Collections:
Conference Paper: From Synthetic to One-Shot Regression of Camera-Agnostic Human Performances
Title | From Synthetic to One-Shot Regression of Camera-Agnostic Human Performances |
---|---|
Authors | |
Keywords | Human performance Monocular video Synthetic data |
Issue Date | 2022 |
Publisher | Springer. |
Citation | Third International Conference on Pattern Recognition and Artificial Intelligence (ICPRAI), Paris, France, June 1–3, 2022. In Pattern Recognition and Artificial Intelligence:Third International Conference, ICPRAI 2022, Paris, France, June 1–3, 2022. Proceedings, Part I, p. 514-525 How to Cite? |
Abstract | Capturing accurate 3D human performances in global space from a static monocular video is an ill-posed problem. It requires solving various depth ambiguities and information about the camera’s intrinsics and extrinsics. Therefore, most methods either learn on given cameras or require to know the camera’s parameters. We instead show that a camera’s extrinsics and intrinsics can be regressed jointly with human’s position in global space, joint angles and body shape only from long sequences of 2D motion estimates. We exploit a static camera’s constant parameters by training a model that can be applied to sequences with arbitrary length with only a single forward pass while allowing full bidirectional information flow. We show that full temporal information flow is especially necessary when improving consistency through an adversarial network. Our training dataset is exclusively synthetic, and no domain adaptation is used. We achieve one of the best Human3.6M joint’s error performances for models that do not use the Human3.6M training data. |
Description | Lecture Notes in Computer Science book series ; volume 13363 |
Persistent Identifier | http://hdl.handle.net/10722/321056 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Habekost, J | - |
dc.contributor.author | Pang, K | - |
dc.contributor.author | Shiratori, T | - |
dc.contributor.author | Komura, T | - |
dc.date.accessioned | 2022-11-01T04:46:09Z | - |
dc.date.available | 2022-11-01T04:46:09Z | - |
dc.date.issued | 2022 | - |
dc.identifier.citation | Third International Conference on Pattern Recognition and Artificial Intelligence (ICPRAI), Paris, France, June 1–3, 2022. In Pattern Recognition and Artificial Intelligence:Third International Conference, ICPRAI 2022, Paris, France, June 1–3, 2022. Proceedings, Part I, p. 514-525 | - |
dc.identifier.uri | http://hdl.handle.net/10722/321056 | - |
dc.description | Lecture Notes in Computer Science book series ; volume 13363 | - |
dc.description.abstract | Capturing accurate 3D human performances in global space from a static monocular video is an ill-posed problem. It requires solving various depth ambiguities and information about the camera’s intrinsics and extrinsics. Therefore, most methods either learn on given cameras or require to know the camera’s parameters. We instead show that a camera’s extrinsics and intrinsics can be regressed jointly with human’s position in global space, joint angles and body shape only from long sequences of 2D motion estimates. We exploit a static camera’s constant parameters by training a model that can be applied to sequences with arbitrary length with only a single forward pass while allowing full bidirectional information flow. We show that full temporal information flow is especially necessary when improving consistency through an adversarial network. Our training dataset is exclusively synthetic, and no domain adaptation is used. We achieve one of the best Human3.6M joint’s error performances for models that do not use the Human3.6M training data. | - |
dc.language | eng | - |
dc.publisher | Springer. | - |
dc.relation.ispartof | Pattern Recognition and Artificial Intelligence:Third International Conference, ICPRAI 2022, Paris, France, June 1–3, 2022. Proceedings, Part I | - |
dc.subject | Human performance | - |
dc.subject | Monocular video | - |
dc.subject | Synthetic data | - |
dc.title | From Synthetic to One-Shot Regression of Camera-Agnostic Human Performances | - |
dc.type | Conference_Paper | - |
dc.identifier.email | Komura, T: taku@cs.hku.hk | - |
dc.identifier.authority | Komura, T=rp02741 | - |
dc.identifier.doi | 10.1007/978-3-031-09037-0_42 | - |
dc.identifier.hkuros | 340628 | - |
dc.identifier.spage | 514 | - |
dc.identifier.epage | 525 | - |
dc.publisher.place | Cham, Germany | - |