TRAINING ADAPTIVE AND SAMPLE-EFFICIENT AUTONOMOUS AGENTS

Loading...
Thumbnail Image
Degree type
PhD
Graduate group
Electrical and Systems Engineering
Discipline
Data Science
Electrical Engineering
Subject
AI Agents
Embodied Agents
Imitation Learning
In-Context Learning
Reinforcement Learning
Funder
Grant number
License
Copyright date
01/01/2025
Distributor
Related resources
Author
Sridhar, Kaustubh
Contributor
Abstract

AI agents, both in the physical and digital worlds, should generalize from their training data to three increasingly difficult levels of deployment: training tasks and environments, training tasks and environments with variations, and completely new tasks and environments. Moreover, like humans, they are expected to learn from as little training data as possible, especially in the physical world, and adapt with as little adaptation data as possible. This thesis is founded around and describes work that tackles these levels of generalization with an additional emphasis on sample-efficiency. We start with a focus on training data efficiency and the simplest level of generalization from training data to training tasks and environments (a.k.a., level 1). AI agents, especially in the physical world, are usually trained via one of two paradigms: imitation learning or reinforcement learning. First, we propose a plug-in model class to improve behavior cloning with any deep neural network (DNN) backbone that is particularly effective in the low-data regime. Second, we leverage our proposed model class to guarantee the conformance of any DNN world model to physics and medical constraints, in a highly data-efficient manner. Third, we improve the sample-efficiency of reinforcement learning agents, by an order of magnitude, by leveraging expert interventions. Next, we tackle the challenge of generalization to training tasks and environments with variations as well as completely new tasks and environments (a.k.a., levels 2 and 3), keeping both training and adaptation sample-efficiency in mind. Here, we pre-train REGENT, a retrieval-augmented generalist agent that can adapt to unseen robotics and game-playing environments via in-context learning, without any finetuning. REGENT outperforms state-of-the-art generalist agents after pre-training on an order-of-magnitude fewer datapoints and with up to 3x fewer parameters. We also propose a strategy, inspired by adaptive control, to improve the robustness of the image encoder of REGENT, an essential component for handling environment variations. Finally, we bring REGENT to the real world by adding in-context adaptability to a pre-trained Vision Language Action model (VLA). Our method, called RICL, creates a RICL-VLA capable of adapting to completely new tasks with unseen objects and novel motions through retrieval-augmentation and in-context learning. Further task-specific RICL finetuning boosts performance, significantly improving over the base VLA directly finetuned on the same data. We conclude by outlining future directions to expand the envelope of tasks and environments to which a general AI agent can adapt.

Advisor
Lee, Insup
Date of degree
2025
Date Range for Data Collection (Start Date)
Date Range for Data Collection (End Date)
Digital Object Identifier
Series name and number
Volume number
Issue number
Publisher
Publisher DOI
Journal Issue
Comments
Recommended citation