Action recognition with the augmented MoCap data using neural data translation

Shih Yao Lin, Yen-Yu Lin

研究成果: Paper同行評審

摘要

This study aims at generating reliable augmented training data to learn a robust deep model for action recognition. The prior knowledge inferred from few training data is not sufficient to well represent the real data distribution, which makes action recognition quite challenging. Inspired by the recent advances in neural machine translation, we propose a neural data translation (NDT) to tackle the aforementioned issue by directly learning the mapping between paired data of the same action class in an end-to-end fashion. The proposed NDT is a sequence-to-sequence generative model. It can be trained with few paired training data, and generates an abundant set of augmented actions with diverse appearance. Specifically, we adopt stochastic pair selection to compile a set of paired training data. Each pair consists of two actions of the same class. One action serves as the input to NDT, while the other acts as the desired output. By learning the mapping between data of the same class, NDT implicitly encodes the intra-class variations so that it can synthesize high-quality actions for augmentation. We evaluated our method on two public datasets, including the Florence3D-action and UCI HAR datasets. The promising results demonstrate that the actions generated by our method effectively improve the performance of action recognition with few examples.

原文English
出版狀態Published - 3 9月 2018
事件29th British Machine Vision Conference, BMVC 2018 - Newcastle, United Kingdom
持續時間: 3 9月 20186 9月 2018

Conference

Conference29th British Machine Vision Conference, BMVC 2018
國家/地區United Kingdom
城市Newcastle
期間3/09/186/09/18

指紋

深入研究「Action recognition with the augmented MoCap data using neural data translation」主題。共同形成了獨特的指紋。

引用此