Abstract | ||
---|---|---|
We present a universal visual navigation method which allows a vehicle to autonomously repeat paths previously taught by a human operator. The method is computationally efficient and does not require camera calibration. It can learn and autonomously traverse arbitrarily shaped paths and is robust to appearance changes induced by varying outdoor illumination and naturally-occurring environment changes. The method does not perform explicit position estimation in the 2d/3d space, but it relies on a novel mathematical theorem, which allows fusing exteroceptive and interoceptive sensory data in a way that ensures navigation accuracy and reliability. The experiments performed indicate that the proposed navigation method can accurately guide different autonomous vehicles along the desired path. The presented system, which was already deployed in patrolling scenarios, is provided as open source at www.github.com/gestom/stroll_bearnav. |
Year | DOI | Venue |
---|---|---|
2018 | 10.1007/978-3-030-14984-0_8 | Lecture Notes in Computer Science |
Field | DocType | Volume |
Computer vision,Human operator,Computer science,Patrolling,Visual navigation,Camera resectioning,Artificial intelligence,Traverse | Conference | 11472 |
ISSN | Citations | PageRank |
0302-9743 | 0 | 0.34 |
References | Authors | |
0 | 9 |
Name | Order | Citations | PageRank |
---|---|---|---|
Filip Majer | 1 | 2 | 3.15 |
Lucie Halodova | 2 | 0 | 1.69 |
Tomas Vintr | 3 | 7 | 3.63 |
Martin Dlouhý | 4 | 0 | 0.34 |
Lukás Merenda | 5 | 0 | 0.34 |
Jaime Pulido | 6 | 63 | 6.23 |
David Portugal | 7 | 175 | 18.74 |
Micael S. Couceiro | 8 | 337 | 28.66 |
Tomás Krajník | 9 | 422 | 37.83 |