Title
ManipulaTHOR: A Framework for Visual Object Manipulation
Abstract
The domain of Embodied AI has recently witnessed substantial progress, particularly in navigating agents within their environments. These early successes have laid the building blocks for the community to tackle tasks that require agents to actively interact with objects in their environment. Object manipulation is an established research domain within the robotics community and poses several challenges including manipulator motion, grasping and long-horizon planning, particularly when dealing with oft-overlooked practical setups involving visually rich and complex scenes, manipulation using mobile agents (as opposed to tabletop manipulation), and generalization to unseen environments and objects. We propose a framework for object manipulation built upon the physics-enabled, visually rich An-THOR framework and present a new challenge to the Embodied AI community known as ArmPoint-Nay. This task extends the popular point navigation task [2] to object manipulation and offers new challenges including 3D obstacle avoidance, manipulating objects in the presence of occlusion, and multi-object manipulation that necessitates long term planning. Popular learning paradigms that are successful on PointNav challenges show promise, but leave a large room for improvement.
Year
DOI
Venue
2021
10.1109/CVPR46437.2021.00447
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021
DocType
ISSN
Citations 
Conference
1063-6919
0
PageRank 
References 
Authors
0.34
0
8
Name
Order
Citations
PageRank
K. Ehsani143.42
Han Winson221.40
Alvaro Herrasti362.91
Eli VanderBilt400.34
Luca Weihs501.69
Eric Kolve61726.33
Aniruddha Kembhavi752831.87
Roozbeh Mottaghi875042.64