Person Tracking by Fusing Posture Data from UAV Video and Wearable Sensors

Alisher Mukashev*, Lan Da Van, Susanta Sharma, M. Farhan Tandia, Yu Chee Tseng

*此作品的通信作者

研究成果: Article同行評審

3 引文 斯高帕斯(Scopus)

摘要

In this article, a novel framework that fuses the posture data taken by a drone (or unmanned aerial vehicle, UAV) camera and the wearable sensors data recorded by smartwatches is proposed. The framework is designed for continuously tracking persons in a drone view by analyzing location-independent human posture features and correctly tagging smartwatch identities (IDs) and personal profiles to video human objects, thus conquering the former work in requiring ground markers. Person detection, ID assignment, and pose estimation are integrated into our framework to obtain recognized human postures. These recognized postures are then paired with those from the wearable sensors. Through fusing common postures, such as standing, walking, jumping, and falling down, person tracking accuracy by UAV up to 95.36% can be attained in our testing scenarios.

原文English
頁(從 - 到)24150-24160
頁數11
期刊IEEE Sensors Journal
22
發行號24
DOIs
出版狀態Published - 15 12月 2022

指紋

深入研究「Person Tracking by Fusing Posture Data from UAV Video and Wearable Sensors」主題。共同形成了獨特的指紋。

引用此