English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT
  Single-Shot Multi-Person 3D Body Pose Estimation From Monocular RGB Input

Mehta, D., Sotnychenko, O., Mueller, F., Xu, W., Sridhar, S., Pons-Moll, G., et al. (2017). Single-Shot Multi-Person 3D Body Pose Estimation From Monocular RGB Input. Retrieved from http://arxiv.org/abs/1712.03453.

Item is

Basic

show hide
Genre: Paper
Latex : Single-Shot Multi-Person {3D} Body Pose Estimation From Monocular {RGB} Input

Files

show Files
hide Files
:
arXiv:1712.03453.pdf (Preprint), 8MB
Name:
arXiv:1712.03453.pdf
Description:
File downloaded from arXiv at 2018-02-01 10:36
OA-Status:
Visibility:
Public
MIME-Type / Checksum:
application/pdf / [MD5]
Technical Metadata:
Copyright Date:
-
Copyright Info:
-

Locators

show

Creators

show
hide
 Creators:
Mehta, Dushyant1, Author           
Sotnychenko, Oleksandr1, Author           
Mueller, Franziska1, Author           
Xu, Weipeng1, Author           
Sridhar, Srinath2, Author
Pons-Moll, Gerard3, Author           
Theobalt, Christian1, Author           
Affiliations:
1Computer Graphics, MPI for Informatics, Max Planck Society, ou_40047              
2External Organizations, ou_persistent22              
3Computer Vision and Multimodal Computing, MPI for Informatics, Max Planck Society, ou_1116547              

Content

show
hide
Free keywords: Computer Science, Computer Vision and Pattern Recognition, cs.CV
 Abstract: We propose a new efficient single-shot method for multi-person 3D pose estimation in general scenes from a monocular RGB camera. Our fully convolutional DNN-based approach jointly infers 2D and 3D joint locations on the basis of an extended 3D location map supported by body part associations. This new formulation enables the readout of full body poses at a subset of visible joints without the need for explicit bounding box tracking. It therefore succeeds even under strong partial body occlusions by other people and objects in the scene. We also contribute the first training data set showing real images of sophisticated multi-person interactions and occlusions. To this end, we leverage multi-view video-based performance capture of individual people for ground truth annotation and a new image compositing for user-controlled synthesis of large corpora of real multi-person images. We also propose a new video-recorded multi-person test set with ground truth 3D annotations. Our method achieves state-of-the-art performance on challenging multi-person scenes.

Details

show
hide
Language(s): eng - English
 Dates: 2017-12-092017
 Publication Status: Published online
 Pages: 11 p.
 Publishing info: -
 Table of Contents: -
 Rev. Type: -
 Identifiers: arXiv: 1712.03453
URI: http://arxiv.org/abs/1712.03453
BibTex Citekey: Mehta1712.03453
 Degree: -

Event

show

Legal Case

show

Project information

show

Source

show