TY - GEN
T1 - TrackNet: A Deep Learning Network for Tracking High-speed and Tiny Objects in Sports Applications
AU - Huang, Yu Chuan
AU - Liao, I. No
AU - Chen, Ching Hsuan
AU - Ik, Tsi-Ui
AU - Peng, Wen-Chih
N1 - Publisher Copyright:
© 2019 IEEE.
PY - 2019/9/18
Y1 - 2019/9/18
N2 - Ball trajectory data are one of the most fundamental and useful information in the evaluation of players' performance and analysis of game strategies. It is still challenging to recognize and position a high-speed and tiny ball accurately from an ordinary video. In this paper, we develop a deep learning network, called TrackNet, to track the tennis ball from broadcast videos in which the ball images are small, blurry, and sometimes with afterimage tracks or even invisible. The proposed heatmap-based deep learning network is trained to not only recognize the ball image from a single frame but also learn flying patterns from consecutive frames. The network is evaluated on the video of the men's singles final at the 2017 Summer Universiade, which is available on YouTube. The precision, recall, and F1 -measure reach 99.7%, 97.3%, and 98.5%, respectively. To prevent overfitting, 9 additional videos are partially labeled together with a subset from the previous dataset to implement 10-fold cross-validation, and the precision, recall, and F1 -measure are 95.3%, 75.7%, and 84.3%, respectively.
AB - Ball trajectory data are one of the most fundamental and useful information in the evaluation of players' performance and analysis of game strategies. It is still challenging to recognize and position a high-speed and tiny ball accurately from an ordinary video. In this paper, we develop a deep learning network, called TrackNet, to track the tennis ball from broadcast videos in which the ball images are small, blurry, and sometimes with afterimage tracks or even invisible. The proposed heatmap-based deep learning network is trained to not only recognize the ball image from a single frame but also learn flying patterns from consecutive frames. The network is evaluated on the video of the men's singles final at the 2017 Summer Universiade, which is available on YouTube. The precision, recall, and F1 -measure reach 99.7%, 97.3%, and 98.5%, respectively. To prevent overfitting, 9 additional videos are partially labeled together with a subset from the previous dataset to implement 10-fold cross-validation, and the precision, recall, and F1 -measure are 95.3%, 75.7%, and 84.3%, respectively.
UR - http://www.scopus.com/inward/record.url?scp=85076366578&partnerID=8YFLogxK
U2 - 10.1109/AVSS.2019.8909871
DO - 10.1109/AVSS.2019.8909871
M3 - Conference contribution
AN - SCOPUS:85076366578
T3 - 2019 16th IEEE International Conference on Advanced Video and Signal Based Surveillance, AVSS 2019
BT - 16th IEEE International Workshop of Content-Aware Video Analysis (CAVA 2019) in conjunction with the 16th IEEE International Conference on Advanced Video and Signal-based Surveillance (AVSS 2019)
PB - IEEE
T2 - 16th IEEE International Conference on Advanced Video and Signal Based Surveillance, AVSS 2019
Y2 - 18 September 2019 through 21 September 2019
ER -