Semantically Synchronizing Multiple-Camera Systems with Human Pose Estimation

Sensors (Basel). 2021 Apr 2;21(7):2464. doi: 10.3390/s21072464.

Abstract

Multiple-camera systems can expand coverage and mitigate occlusion problems. However, temporal synchronization remains a problem for budget cameras and capture devices. We propose an out-of-the-box framework to temporally synchronize multiple cameras using semantic human pose estimation from the videos. Human pose predictions are obtained with an out-of-the-shelf pose estimator for each camera. Our method firstly calibrates each pair of cameras by minimizing an energy function related to epipolar distances. We also propose a simple yet effective multiple-person association algorithm across cameras and a score-regularized energy function for improved performance. Secondly, we integrate the synchronized camera pairs into a graph and derive the optimal temporal displacement configuration for the multiple-camera system. We evaluate our method on four public benchmark datasets and demonstrate robust sub-frame synchronization accuracy on all of them.

Keywords: epipolar geometry; human pose estimation; multiple-camera system; temporal synchronization.

MeSH terms

  • Algorithms*
  • Humans