FlickerNet: Adaptive 3D Gesture Recognition from Sparse Point Clouds

Yuecong Min (Institute of Computing Technology, Chinese Academy of Sciences), Xiujuan Chai (Agricultural Information Institute), Lei Zhao (HUAWEI Technologies Co., Ltd.), Xilin Chen (Institute of Computing Technology, Chinese Academy of Sciences)

Abstract
Recent studies on gesture recognition use deep convolutional neural networks (CNNs) to extract spatiotemporal features from individual frames or short video clips. However, extracting features frame-by-frame will bring a lot of redundant and ambiguous gesture information. Inspired by the flicker fusion phenomena, we propose a simple but efficient network, called FlickerNet, to recognize gesture from a sequence of sparse point clouds sampled from depth videos. Different from the existing CNN-based methods, FlickerNet can adaptively recognize hand postures and hand motions from the flicker of gestures: the point clouds of the stable hand postures and the sparse point-cloud motion for fast hand motions. Notably, FlickerNet significantly outperforms the previous state-of-the-art approaches on two challenging datasets with much higher computational efficiency.

DOI
10.5244/C.33.57
https://dx.doi.org/10.5244/C.33.57

Files
Paper (PDF)
Supplementary material (ZIP)

BibTeX
@inproceedings{BMVC2019,
title={FlickerNet: Adaptive 3D Gesture Recognition from Sparse Point Clouds},
author={Yuecong Min and Xiujuan Chai and Lei Zhao and Xilin Chen},
year={2019},
month={September},
pages={57.1--57.13},
articleno={57},
numpages={13},
booktitle={Proceedings of the British Machine Vision Conference (BMVC)},
publisher={BMVA Press},
editor={Kirill Sidorov and Yulia Hicks},
doi={10.5244/C.33.57},
url={https://dx.doi.org/10.5244/C.33.57}
}