Abstract | ||
---|---|---|
We developed an exploratory VR environment, where spatial features and narratives can be manipulated in real time by the facial and head gestures of the user. We are using the Faceteq prototype, exhibited in 2017, as the interactive interface. Faceteq consists of a wearable technology that can be adjusted on commercial HMDs for measuring facial expressions and biometric responses. Faceteq project was founded with the aim to provide a human-centred additional tool for affective human-computer interaction. The proposed demo will exhibit the hardware and the functionality of the demo in real time.
|
Year | DOI | Venue |
---|---|---|
2017 | 10.1145/3131277.3134366 | SUI '17: Symposium on Spatial User Interaction
Brighton
United Kingdom
October, 2017 |
Keywords | Field | DocType |
Virtual Reality,Facial Expression,Emotion,EMG,Affective Computing | Virtual reality,Gesture,Computer science,Human–computer interaction,Artificial intelligence,Wearable technology,Computer vision,Narrative,Facial expression,Biometrics,Affective computing,Affect (psychology),Multimedia | Conference |
ISBN | Citations | PageRank |
978-1-4503-5486-8 | 0 | 0.34 |
References | Authors | |
0 | 8 |
Name | Order | Citations | PageRank |
---|---|---|---|
Ifigeneia Mavridou | 1 | 0 | 0.34 |
Mahyar Hamedi | 2 | 18 | 3.84 |
Mohsen Fatoorechi | 3 | 0 | 1.01 |
J. Archer | 4 | 0 | 0.34 |
Andrew Cleal | 5 | 0 | 0.68 |
Emili Balaguer-Ballester | 6 | 59 | 10.97 |
E Seiss | 7 | 15 | 4.34 |
Charles Nduka | 8 | 4 | 2.77 |