Task-Unaware Lifelong Robot Learning with Retrieval-based Weighted Local Adaptation
P. Yang (TU Delft - Electrical Engineering, Mathematics and Computer Science)
C. Wang – Mentor (TU Delft - Learning & Autonomous Control)
J. Kober – Mentor (TU Delft - Learning & Autonomous Control)
Frans Oliehoek – Mentor (TU Delft - Sequential Decision Making)
Chirag Raman – Graduation committee member (TU Delft - Pattern Recognition and Bioinformatics)
More Info
expand_more
Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.
Abstract
Real-world environments require robots to continuously acquire new skills while retain-ing previously learned abilities, all without the need for clearly defined task boundaries. Storing all past data to prevent forgetting is impractical due to storage and privacy con-cerns. To address this, we propose a method that efficiently restores a robot’s proficiency in previously learned tasks over its lifespan. Using an Episodic Memory M, our approach enables experience replay during training and retrieval during testing for local fine-tuning, allowing rapid adaptation to previously encountered problems without explicit task iden-tifiers. Additionally, we introduce a selective weighting mechanism that emphasizes the most challenging segments of retrieved demonstrations, focusing local adaptation where it is most needed. This framework offers a scalable solution for lifelong learning in dy-namic, task-unaware environments, combining retrieval-based local adaptation with selec-tive weighting to enhance robot performance in open-ended scenarios.