Paper-Reading

EgoCap: Egocentric Marker-less Motion Capture with Two Fisheye Cameras

tags: Deep Learning, Motion Capture, Pose Estimation, SIGGRAPH 2016

Summary:

New method for real-time, marker-less, and egocentric motion capture:

estimating the full-body skeleton pose from a lightweight stereo pair of fisheye cameras attached to a helmet or virtual reality headset。

EgoCap: an egocentric motion-capture approach that estimates full-body pose from a pair of optical cameras carried by lightweight headgear

Contribution

  1. a new egocentric inside-in sensor rig with only two head-mounted, downward-facing commodity video cameras with fisheye lenses
  2. a new marker-less motion capture algorithm tailored to the strongly distorted egocentric fisheye views.
    • combines a generative model-based skeletal pose estimation approach (Section 4) with evidence from a trained ConvNet-based body part detector (Section 4.3)
    • features an analytically differentiable objective energy that can be minimized efficiently
    • work with unsegmented frames and general backgrounds
    • succeeds even on poses exhibiting notable self-occlusions
    • part detector predicts occluded parts, and enables recovery from tracking errors after severe occlusions.
  3. a new approach for automatically creating body part detection training datasets

Egocentric Camera Design

Egocentric Camera Design

Egocentric Full-Body Motion Capture

The setup separates human motion capture into two sub-problems:

  1. local skeleton pose estimation with respect to the camera rig
  2. global rig pose estimation relative to the environment

4.1 Body Model

4.2 Egocentric Volumetric Ray-Casting Model

4.3 Egocentric Body-Part Detection

4.3.1 Dataset Creation

Dataset Augmentation

4.3.2 Detector Learning

4.3.3 Body-Part Detection Energy

4.4 Real-Time Optimization

Evaluation

5.3 Body-Part Detections

Dataset Augmentations

Detection Accuracy

5.4 3D Body Pose Accuracy

##Applications

6.3 Tracking for Immersive VR

Global Pose Estimation