Paper: | WA-L2.3 |
Session: | Video Object Segmentation and Tracking II |
Time: | Wednesday, September 19, 10:30 - 10:50 |
Presentation: |
Lecture
|
Title: |
MONOCULAR TRACKING 3D PEOPLE BY GAUSSIAN PROCESS SPATIO-TEMPORAL VARIABLE MODEL |
Authors: |
Junbiao Pang; Institute of Computing Technology, Chinese Academy of Sciences | | |
| Laiyun Qing; Institute of Computing Technology, Chinese Academy of Sciences | | |
| Qingming Huang; Institute of Computing Technology, Chinese Academy of Sciences | | |
| Shuqiang Jiang; Institute of Computing Technology, Chinese Academy of Sciences | | |
| Wen Gao; Institute of Digital Media, Peking University | | |
Abstract: |
Tracking 3D people from monocular video is often poorly constrained. To mitigate this problem, prior knowledge should be exploited. In this paper, the Gaussian process spatio-temporal variable model (GPSTVM), a novel dynamical system modeling method is proposed for learning human pose and motion priors. The GPSTVM provides a low dimensional embedding of human motion data, with a smooth density function that provides higher probability to the poses and motions close to the training data. The low dimensional latent space is optimized directly to retain the spatio-temporal structure of the high dimensional pose space. After the prior on human pose is learned, the particle filtering can be used tracking articulated human pose; particle filtering propagates over time in the embedding space, avoiding the curse of dimensionality. Experiments demonstrate that our approach tracks 3D people accurately. |