@inbook{deffe70b86be49ee88dfc83e27a76980,
title = "Learning and inferring human actions with temporal pyramid features based on conditional random fields",
abstract = "Finding an effective way to represent human actions is yet an open problem because it usually requires taking evidences extracted from various temporal resolutions into account. A conventional way of representing an action employs tem-porally ordered fine-grained movements, e.g., key poses or subtle motions. Many existing approaches model actions by directly learning the transitional relationships between those fine-grained features. Yet, an action data may have many similar observations with occasional and irregular changes, which make commonly used fine-grained features less reli-able. This paper presents a set of temporal pyramid features that enriches action representation with various levels of se-mantic granularities. For learning and inferring the proposed pyramid features, we adopt a discriminative model with latent variables to capture the hidden dynamics in each layer of the pyramid. Our method is evaluated on a Tai-Chi Chun dataset and a daily activities dataset. Both of them are collected by us. Experimental results demonstrate that our approach achieves more favorable performance than existing methods.",
keywords = "conditional random fields, human action recognition, temporal pyramid representation",
author = "Lin, {Shih Yao} and Lin, {Yen Yu} and Chen, {Chu Song} and Hung, {Yi Ping}",
year = "2017",
month = jun,
day = "16",
doi = "10.1109/ICASSP.2017.7952630",
language = "American English",
isbn = "9781509041176",
series = "ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
pages = "2617--2621",
booktitle = "2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)",
address = "United States",
}