Abstract | ||
---|---|---|
Machine Learning (ML) models are widely used to infer human activities. However, collecting data to train ML models in realworld often requires significant time and effort. In this paper, we suggest a novel data collection framework to leverage pre-built VR applications and simulating tools. In particular, we applied the concept of virtual Inertial Measurement Unit (IMU) to capture activities of an avatar in simulation. Our initial results show that Random Forest (RF), Support Vector Machine (SVM), and Long Short Term Memory (LSTM) models built with the virtual sensor data can classify three activities (i.e., standing, running, walking) over a realworld dataset at the accuracy of 80.40% (87.83% precision and 80.12% recall), 67.52% accuracy (72.24% precision and 68.15% recall), and 77.67% accuracy (86.25% precision and 77.63% recall), respectively. The early results show the initial feasibility of simulation-driven machine learning without real-world data
|
Year | DOI | Venue |
---|---|---|
2019 | 10.1145/3325424.3329662 | The 5th ACM Workshop on Wearable Systems and Applications |
Keywords | Field | DocType |
activity recognition, imu, machine learning, sensor, simulation, virtual reality | Computer science,Artificial intelligence,Machine learning | Conference |
ISBN | Citations | PageRank |
978-1-4503-6775-2 | 0 | 0.34 |
References | Authors | |
0 | 3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Cholmin Kang | 1 | 0 | 0.34 |
Hyunwoo Jung | 2 | 0 | 0.68 |
Youngki Lee | 3 | 832 | 70.33 |