Abstract | ||
---|---|---|
In-the-wild human pose estimation has a huge - potential for various fields, ranging from animation and action recognition to intention recognition and prediction for autonomous driving. The current state-of-the-art is focused only on RGB and RGB-D approaches for predicting the 3D human pose. However, not using precise LiDAR depth information limits the performance and leads to very inaccurate absolute pose estimation. With LiDAR sensors becoming more affordable and common on robots and autonomous vehicle setups, we propose an end-to-end architecture using RGB and LiDAR to predict the absolute 3D human pose with unprecedented precision. Additionally, we introduce a weakly-supervised approach to generate 3D predictions using 2D pose annotations from PedX This allows for many new opportunities in the field of 3D human pose estimation. |
Year | DOI | Venue |
---|---|---|
2020 | 10.1109/ICPR48806.2021.9412785 | 2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR) |
Keywords | DocType | ISSN |
sensor fusion, 3D human pose estimation, Li-DAR, RGB, autonomous vehicles, perception | Conference | 1051-4651 |
Citations | PageRank | References |
0 | 0.34 | 0 |
Authors | ||
5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Michael Fürst | 1 | 0 | 0.34 |
Shriya T. P. Gupta | 2 | 0 | 0.34 |
René Schuster | 3 | 14 | 5.44 |
Oliver Wasenmüller | 4 | 21 | 6.24 |
Didier Stricker | 5 | 1266 | 138.03 |