Multi-Person Monocular 3D Pose Estimation Using Deep Neural Networks

Thumbnail

Event details

Date 27.06.2017
Hour 10:3012:30
Speaker Isinsu Katircioglu
Location
Category Conferences - Seminars

EDIC candidacy exam
Exam president: Prof. Mark Pauly
Thesis advisor: Prof. Pascal Fua
Thesis co-advisor: Dr. Mathieu Salzmann
Co-examiner: Prof. Sabine Süsstrunk

Abstract
Recovering 3D pose of multiple interacting subjects in unconstrained environments is highly challenging due to strong partial occlusions and ambiguities caused by the close proximity of individuals. In this work, we address the problem of multi- person 3D pose estimation from monocular RGB images using deep neural networks. First we discuss three related methods that tackle different aspects of our problem: a framework for 3D human pose prediction using shape context matching, a deep learning model for multi-person 2D pose estimation that is cast as an integer linear programming problem and an approach for identifying collective activities from videos containing a group of interacting people. Finally, we introduce our deep learning method that regresses individual 3D poses from heatmaps of 2D joint locations and feeds them to a dependency network to model the correlations between them. We evaluate the performance of our approach on a new boxing dataset.
Background papers

DeeperCut: A Deeper, Stronger, and Faster Multi-Person Pose Estimation Model. In ECCV'16, by  Insafutdinov, E., Pishchulin, L., Andres, B., Andriluka, M., Schiele, B.
Understanding Collective Activities of People from Videos. In PAMI'14, by  Choi, W., Savarese S.
Recovering 3D Human Body Configurations Using Shape Contexts. In PAMI'2006, by Mori, G., Malik, J.

 

Practical information

  • General public
  • Free

Contact

Tags

EDIC candidacy exam

Share