Abstract | ||
---|---|---|
We present an efficient approach for real-time continuous human action recognition with depth sensors. Instead of using the powerful but quite complex deep neural networks, our approach uses a light-weight discriminative frame-level descriptor, which is called the context pose (CP). The objective is to make our approach realistic in mobile depth sensor applications. CP integrates the context of the motion within a depth video. CP can increase the discriminative power of the frames and enable more frames to represent actions. CP is incorporated with a new part-based random decision forest (PRDF) method. The PRDF is designed to automatically select the optimal combination of body parts to represent and distinguish each action. We evaluate our approach on three classical single-action benchmark datasets. The experiments show that our approach has 200 frames/s and wins superior performances to the existing frame-level descriptors and classifiers in terms of accuracy. |
Year | DOI | Venue |
---|---|---|
2018 | 10.1109/ACCESS.2018.2869330 | IEEE ACCESS |
Keywords | Field | DocType |
Depth sensor, sensor systems and applications, pattern recognition | Computer science,Action recognition,Artificial intelligence,Artificial neural network,Hidden Markov model,Random forest,Discriminative model,Deep neural networks,Machine learning,Distributed computing | Journal |
Volume | ISSN | Citations |
6 | 2169-3536 | 0 |
PageRank | References | Authors |
0.34 | 0 | 7 |
Name | Order | Citations | PageRank |
---|---|---|---|
Hejun Wu | 1 | 242 | 23.03 |
Zhenye Huang | 2 | 0 | 0.34 |
Biao Hu | 3 | 1 | 0.69 |
Zhi Yu | 4 | 1 | 2.38 |
Xiying Li | 5 | 3 | 3.81 |
Min Gao | 6 | 17 | 6.08 |
Zhong Shen | 7 | 8 | 4.64 |