Abstract | ||
---|---|---|
An exoskeleton haptic interface is developed for functional training in virtual environments. A composite control scheme enables a variety of tasks to be implemented, and a "Qt" graphics library is used to generate the virtual environment for the haptic interface at the hand and graphical user interfaces for input and telemetry. Inter-process communications convert telemetry from the exoskeleton into motion commands for objects in the virtual environment. A second haptic interface at the upper arm is used to control the elbow orbit self-motion of the arm during tasks. Preliminary results are reviewed for a wall-painting task in which the virtual wall stiffness and viscosity are generated using an admittance controller. |
Year | DOI | Venue |
---|---|---|
2009 | 10.1109/IROS.2009.5354834 | IROS |
Keywords | Field | DocType |
force,virtual reality,exoskeletons,graphical user interfaces,functional training,haptic interface,telemetry,inter process communication,graphic user interface,virtual environment,human robot interaction | Control theory,Virtual machine,Virtual reality,Computer science,Simulation,Graphical user interface,Human–computer interaction,Exoskeleton,Admittance,Haptic technology,Human–robot interaction | Conference |
Citations | PageRank | References |
17 | 1.30 | 7 |
Authors | ||
3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Craig R. Carignan | 1 | 83 | 12.38 |
Jonathan Tang | 2 | 17 | 1.30 |
Stephen Roderick | 3 | 45 | 3.34 |