Learning adaptive hidden layers for mobile gesture recognition

Ting Kuei Hu, Yen Yu Lin, Pi Cheng Hsiu

Research output: Chapter in Book/Report/Conference proceedingChapterpeer-review

13 Scopus citations


This paper addresses two obstacles hindering advances in accurate gesture recognition on mobile devices. First, gesture recognition performance is highly dependant on feature selection, but optimal features typically vary from gesture to gesture. Second, diverse user behaviors and mobile environments result in extremely large intra-class variations. We tackle these issues by introducing a new network layer, called an adaptive hidden layer (AHL), to generalize a hidden layer in deep neural networks and dynamically generate an activation map conditioned on the input. To this end, an AHL is composed of multiple neuron groups and an extra selector. The former compiles multi-modal features captured by mobile sensors, while the latter adaptively picks a plausible group for each input sample. The AHL is end-to-end trainable and can generalize an arbitrary subset of hidden layers. Through a series of AHLs, the great expressive power from exponentially many forward paths allows us to choose proper multi-modal features in a sample-specific fashion and resolve the problems caused by the unfavorable variations in mobile gesture recognition. The proposed approach is evaluated on a benchmark for gesture recognition and a newly collected dataset. Superior performance demonstrates its effectiveness. Copyright © 2018, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved.
Original languageAmerican English
Title of host publication32nd AAAI Conference on Artificial Intelligence, AAAI 2018
PublisherAAAI press
Number of pages9
ISBN (Print)9781577358008
StatePublished - Apr 2018

Publication series

Name32nd AAAI Conference on Artificial Intelligence, AAAI 2018


Dive into the research topics of 'Learning adaptive hidden layers for mobile gesture recognition'. Together they form a unique fingerprint.

Cite this