论文标题
字母:全身区域多人姿势估计和实时跟踪
AlphaPose: Whole-Body Regional Multi-Person Pose Estimation and Tracking in Real-Time
论文作者
论文摘要
准确的全身多人姿势估计和跟踪是计算机视觉中重要但充满挑战的话题。为了捕获人类在复杂行为分析中的微妙动作,在传统的纯体姿势估计中,全身姿势估计包括面部,身体,手和脚是必不可少的。在本文中,我们提出字母,该系统可以在实时运行时执行准确的全身姿势估计和跟踪。为此,我们提出了几种新技术:用于快速和良好本地化的对称积分关键点回归(SIKR),参数姿势非最大抑制(P-NMS),以消除冗余人类检测,并构成姿势认识的身份嵌入,以构成共同姿势估计和跟踪。在培训期间,我们求助于部分引导的提案生成器(PGPG)和多域知识蒸馏,以进一步提高准确性。我们的方法能够准确地定位全身关键点,并同时跟踪人类给出不准确的边界框和冗余检测。我们在Coco-Whole Body,Coco,Posetrack以及我们提出的Halpe-Fulloby姿势估计数据集方面,对当前最新方法的速度和准确性都具有显着改善。我们的模型,源代码和数据集可在https://github.com/mvig-sjtu/alphapose上公开提供。
Accurate whole-body multi-person pose estimation and tracking is an important yet challenging topic in computer vision. To capture the subtle actions of humans for complex behavior analysis, whole-body pose estimation including the face, body, hand and foot is essential over conventional body-only pose estimation. In this paper, we present AlphaPose, a system that can perform accurate whole-body pose estimation and tracking jointly while running in realtime. To this end, we propose several new techniques: Symmetric Integral Keypoint Regression (SIKR) for fast and fine localization, Parametric Pose Non-Maximum-Suppression (P-NMS) for eliminating redundant human detections and Pose Aware Identity Embedding for jointly pose estimation and tracking. During training, we resort to Part-Guided Proposal Generator (PGPG) and multi-domain knowledge distillation to further improve the accuracy. Our method is able to localize whole-body keypoints accurately and tracks humans simultaneously given inaccurate bounding boxes and redundant detections. We show a significant improvement over current state-of-the-art methods in both speed and accuracy on COCO-wholebody, COCO, PoseTrack, and our proposed Halpe-FullBody pose estimation dataset. Our model, source codes and dataset are made publicly available at https://github.com/MVIG-SJTU/AlphaPose.