Person Tracking by Fusing Posture Data from UAV Video and Wearable Sensors

Alisher Mukashev, Lan Da Van, Susanta Sharma, M. Farhan Tandia, Yu Chee Tseng

研究成果: Article同行評審

2 引文 斯高帕斯(Scopus)


In this paper, a novel framework that fuses the posture data taken by a drone (or unmanned aerial vehicle, UAV) camera and the wearable sensors data recorded by smartwatches is proposed. The framework is designed for continuously tracking persons in a drone view by analyzing location-independent human posture features and correctly tagging smartwatch identities (IDs) and personal profiles to video human objects, thus conquering the former work in requiring ground markers. Person detection, ID assignment, and pose estimation are integrated into our framework to obtain recognized human postures. These recognized postures are then paired with those from the wearable sensors. Through fusing common postures such as standing, walking, jumping, and falling down, person tracking accuracy by UAV up to 95.36% can be attained in our testing scenarios.

頁(從 - 到)1
期刊IEEE Sensors Journal
出版狀態Accepted/In press - 2022


深入研究「Person Tracking by Fusing Posture Data from UAV Video and Wearable Sensors」主題。共同形成了獨特的指紋。