Papers
arxiv:2601.09512

CLARE: Continual Learning for Vision-Language-Action Models via Autonomous Adapter Routing and Expansion

Published on Jan 14
· Submitted by
Ralf Römer
on Jan 20
Authors:
,
,

Abstract

CLARE is a parameter-efficient, exemplar-free continual learning framework for vision-language-action models that enables robots to adapt to new tasks while preserving previously learned knowledge through lightweight adapters and dynamic routing mechanisms.

AI-generated summary

To teach robots complex manipulation tasks, it is now a common practice to fine-tune a pre-trained vision-language-action model (VLA) on task-specific data. However, since this recipe updates existing representations, it is unsuitable for long-term operation in the real world, where robots must continually adapt to new tasks and environments while retaining the knowledge they have already acquired. Existing continual learning methods for robotics commonly require storing previous data (exemplars), struggle with long task sequences, or rely on task identifiers for deployment. To address these limitations, we propose CLARE, a general, parameter-efficient framework for exemplar-free continual learning with VLAs. CLARE introduces lightweight modular adapters into selected feedforward layers and autonomously expands the model only where necessary when learning a new task, guided by layer-wise feature similarity. During deployment, an autoencoder-based routing mechanism dynamically activates the most relevant adapters without requiring task labels. Through extensive experiments on the LIBERO benchmark, we show that CLARE achieves high performance on new tasks without catastrophic forgetting of earlier tasks, significantly outperforming even exemplar-based methods. Code and data are available at https://tum-lsy.github.io/clare.

Community

TL;DR

🤖 CLARE enables Vision-Language-Action models to learn new robot tasks without forgetting previous ones — no replay buffers, no task IDs at inference.

  • 🔌 Plug-and-play adapters: Extends PEFT with a new CLARE adapter type
  • 🧠 Smart expansion: Automatically adds new adapter modules only when needed (based on feature similarity)
  • 🎯 Task-free inference: Autoencoder-based routing selects the right adapters without knowing the task
  • 📈 SOTA on LIBERO: Outperforms exemplar-based continual learning methods on long task sequences

Built on 🤗 LeRobot + PEFT.

Sign up or log in to comment

Models citing this paper 2

Datasets citing this paper 10

Browse 10 datasets citing this paper

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2601.09512 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.