Abstract | ||
---|---|---|
This paper proposes an approach to automatically detect dementia from a human face. Although some works have detected dementia from speech and language attributes, there are few studies focusing on facial expression in dementia patients. We recorded the human-agent interaction data of spoken dialogues from 24 participants (12 with dementia and 12 without) and extracted the face features. Our objective was to classify dementia by L1 regularized logistic regression. The facial features and the L1 logistic regression were then used to classify the participants into two groups with 0.82 detection performance, as measured by the areas under the receiver operating characteristic curve. We also identified various contributing features, such as action units, eye gaze, and lip activity. These results demonstrate that our system has the potential to detect dementia from the face through spoken dialog systems and as such, can be of assistance to health care workers.
|
Year | DOI | Venue |
---|---|---|
2019 | 10.1145/3351529.3360654 | Adjunct of the 2019 International Conference on Multimodal Interaction |
Keywords | Field | DocType |
Alzheimer’s disease, Dementia, agents, eye, facial expression | Psychology,Psychiatry,Dementia | Conference |
ISBN | Citations | PageRank |
978-1-4503-6937-4 | 0 | 0.34 |
References | Authors | |
0 | 6 |
Name | Order | Citations | PageRank |
---|---|---|---|
Hiroki Tanaka | 1 | 29 | 11.87 |
Hiroyoshi Adachi | 2 | 0 | 1.01 |
Hiroaki Kazui | 3 | 4 | 2.22 |
Manabu Ikeda | 4 | 1 | 1.98 |
Takashi Kudo | 5 | 15 | 2.92 |
Satoshi Nakamura | 6 | 4 | 4.84 |