TLDR: The research paper ‘OpenMoCap: Rethinking Optical Motion Capture under Real-world Occlusion’ introduces two key innovations to address performance degradation in optical motion capture due to marker occlusions. First, the CMU-Occlu dataset is presented, which realistically simulates marker occlusion patterns using ray tracing, bridging the gap between synthetic and real-world data. Second, the OpenMoCap model is proposed, featuring a novel marker-joint chain inference mechanism that enables robust motion capture by establishing long-range dependencies between markers and joints. Experiments show OpenMoCap significantly outperforms existing methods, especially in high-occlusion scenarios, and the CMU-Occlu dataset improves model generalization for real-world deployment.
Optical motion capture, often referred to as MoCap, is a fundamental technology that underpins advancements in fields like virtual reality, film production, and game development. It works by using multiple infrared cameras to track reflective markers placed on a subject’s body, allowing for the precise digital recording and reconstruction of human motion. However, a significant challenge arises in real-world applications: marker occlusion. This occurs when markers are blocked by body parts, clothing, or environmental objects, preventing cameras from accurately capturing their positions. This leads to severe performance degradation in existing MoCap systems.
Researchers have identified two primary limitations in current motion capture models. Firstly, there’s a notable lack of training datasets that accurately reflect realistic marker occlusion patterns. Existing synthetic datasets often use random occlusions, which don’t match the sustained and patterned occlusions seen in real-world scenarios. This mismatch causes models trained on such data to perform poorly when deployed in practical settings. Secondly, current models often lack effective strategies to capture long-range dependencies among markers. When multiple adjacent markers are occluded simultaneously, methods that rely only on nearby visible markers struggle to infer the correct positions.
To address these critical issues, a team of researchers from Tsinghua University has introduced two significant innovations: the CMU-Occlu dataset and the OpenMoCap model. The CMU-Occlu dataset is a large-scale motion capture dataset designed to accurately reflect real-world marker occlusion characteristics. Unlike previous datasets, CMU-Occlu incorporates ray tracing techniques to realistically simulate how markers become occluded by body parts or obstacles, significantly enhancing the consistency between synthetic training data and actual occlusion patterns. This dataset serves as a crucial benchmark for evaluating robust motion solving.
Alongside the dataset, the researchers propose OpenMoCap, a novel motion-solving model specifically engineered for robust motion capture in environments with significant occlusions. OpenMoCap features a unique multi-stage framework that decouples position estimation from rotation estimation, addressing critical data loss caused by marker occlusions. Its core innovation lies in the Marker-Joint Chain Inference Mechanism. This mechanism recognizes the mutual relationship between markers and joints, treating joints as intermediate nodes to establish long-range spatial constraints among markers. This means that even if a marker is heavily occluded, information can propagate through related joints to help accurately reconstruct its position. This simultaneous optimization of marker and joint positions leads to more precise reconstruction results.
Also Read:
- Enhancing Visual Perception with Decoupled Learning: Introducing DeCLIP
- Enhancing 3D Scene Reconstruction with Integrated Camera and Depth Information
Extensive comparative experiments have demonstrated the effectiveness of these innovations. OpenMoCap consistently outperforms competing methods across diverse scenarios, showing significant reductions in joint position and joint rotation errors, especially under high-occlusion conditions. The CMU-Occlu dataset has also been shown to provide consistent performance improvements for existing methods, highlighting its value in improving model generalization for real-world deployment. The OpenMoCap algorithm has been integrated into the MoSen MoCap system, offering a low-cost solution that eliminates the need for labor-intensive post-processing common in mainstream commercial systems, thereby reducing overall costs and paving the way for broader adoption of motion capture technology. For more details, you can refer to the research paper.


