Abstract | ||
---|---|---|
This paper presents the results of a feasibility study of a deep learning scheme for sign language motion recognition. Capturing the motions used in sign language was conducted using specially designed colored gloves and an optical camera. Deep learning and conventional classification schemes were used for motion recognition, and their results are compared. In a deep learning process each frame of motion data is passed directly to AlexNet for feature extraction. Although the structure of the neural network and optional parameters for deep learning have not been optimized at this stage, it was verified that the accuracy of recognition ranged from 59.6% to 72.3% for twenty-five motions. Though this performance is inferior to that of conventional schemes, it is considered that these results indicate the feasibility of using a deep learning scheme for sign language motion recognition. |
Year | DOI | Venue |
---|---|---|
2018 | 10.1007/978-3-319-93659-8_103 | COMPLEX, INTELLIGENT, AND SOFTWARE INTENSIVE SYSTEMS |
Field | DocType | Volume |
Computer vision,Colored,Motion recognition,Computer science,Classification scheme,Feature extraction,Sign language,Artificial intelligence,Deep learning,Artificial neural network,Distributed computing | Conference | 772 |
ISSN | Citations | PageRank |
2194-5357 | 0 | 0.34 |
References | Authors | |
2 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Kazuki Sakamoto | 1 | 0 | 0.34 |
Eiji Ota | 2 | 0 | 0.34 |
Tatsunori Ozawa | 3 | 1 | 1.11 |
Hiromitsu Nishimura | 4 | 47 | 7.82 |
Hiroshi Tanaka | 5 | 56 | 13.71 |