Autonomous Learning Conference Paper 2023

Efficient Learning of High Level Plans from Play

Arxiv Website
Thumb ticker sm whatsapp image 2022 10 04 at 12.01.09 pm
Autonomous Learning
  • Doctoral Researcher
Thumb ticker sm prof pic
Empirical Inference
  • Doctoral Researcher
Thumb ticker sm georg 2018 crop small
Empirical Inference, Autonomous Learning
Senior Research Scientist

Real-world robotic manipulation tasks remain an elusive challenge, since they involve both fine-grained environment interaction, as well as the ability to plan for long-horizon goals. Although deep reinforcement learning (RL) methods have shown encouraging results when planning end-to-end in high-dimensional environments, they remain fundamentally limited by poor sample efficiency due to inefficient exploration, and by the complexity of credit assignment over long horizons. In this work, we present Efficient Learning of High-Level Plans from Play (ELF-P), a framework for robotic learning that bridges motion planning and deep RL to achieve long-horizon complex manipulation tasks. We leverage task-agnostic play data to learn a discrete behavioral prior over object-centric primitives, modeling their feasibility given the current context. We then design a high-level goal-conditioned policy which (1) uses primitives as building blocks to scaffold complex long-horizon tasks and (2) leverages the behavioral prior to accelerate learning. We demonstrate that ELF-P has significantly better sample efficiency than relevant baselines over multiple realistic manipulation tasks and learns policies that can be easily transferred to physical hardware.

Author(s): Armengol Urpi, N. and Bagatella, M. and Hilliges, O. and Martius, G. and Coros, S.
Links:
Book Title: International Conference on Robotics and Automation
Year: 2023
Bibtex Type: Conference Paper (inproceedings)
State: Accepted
Electronic Archiving: grant_archive
Attachments:

BibTex

@inproceedings{armengol2023elfp,
  title = {Efficient Learning of High Level Plans from Play},
  booktitle = {International Conference on Robotics and Automation},
  abstract = {Real-world robotic manipulation tasks remain an elusive challenge, since they involve both fine-grained environment interaction, as well as the ability to plan for long-horizon goals. Although deep reinforcement learning (RL) methods have shown encouraging results when planning end-to-end in high-dimensional environments, they remain fundamentally limited by poor sample efficiency due to inefficient exploration, and by the complexity of credit assignment over long horizons. In this work, we present Efficient Learning of High-Level Plans from Play (ELF-P), a framework for robotic learning that bridges motion planning and deep RL to achieve long-horizon complex manipulation tasks. We leverage task-agnostic play data to learn a discrete behavioral prior over object-centric primitives, modeling their feasibility given the current context. We then design a high-level goal-conditioned policy which (1) uses primitives as building blocks to scaffold complex long-horizon tasks and (2) leverages the behavioral prior to accelerate learning. We demonstrate that ELF-P has significantly better sample efficiency than relevant baselines over multiple realistic manipulation tasks and learns policies that can be easily transferred to physical hardware. },
  year = {2023},
  slug = {armengol2023elfp},
  author = {Armengol Urpi, N. and Bagatella, M. and Hilliges, O. and Martius, G. and Coros, S.}
}