TRAINING ADAPTIVE AND SAMPLE-EFFICIENT AUTONOMOUS AGENTS

Loading...
Thumbnail Image

Degree type

Doctor of Philosophy (PhD)

Graduate group

Electrical and Systems Engineering

Discipline

Data Science
Electrical Engineering

Subject

AI Agents
Embodied Agents
Imitation Learning
In-Context Learning
Reinforcement Learning

Funder

Grant number

License

Copyright date

2025

Distributor

Related resources

Contributor

Abstract

AI agents, both in the physical and digital worlds, should generalize from their training data to three increasingly difficult levels of deployment: training tasks and environments, training tasks and environments with variations, and completely new tasks and environments. Moreover, like humans, they are expected to learn from as little training data as possible, especially in the physical world, and adapt with as little adaptation data as possible. This thesis is founded around and describes work that tackles these levels of generalization with an additional emphasis on sample-efficiency. We start with a focus on training data efficiency and the simplest level of generalization from training data to training tasks and environments (a.k.a., level 1). AI agents, especially in the physical world, are usually trained via one of two paradigms: imitation learning or reinforcement learning. First, we propose a plug-in model class to improve behavior cloning with any deep neural network (DNN) backbone that is particularly effective in the low-data regime. Second, we leverage our proposed model class to guarantee the conformance of any DNN world model to physics and medical constraints, in a highly data-efficient manner. Third, we improve the sample-efficiency of reinforcement learning agents, by an order of magnitude, by leveraging expert interventions. Next, we tackle the challenge of generalization to training tasks and environments with variations as well as completely new tasks and environments (a.k.a., levels 2 and 3), keeping both training and adaptation sample-efficiency in mind. Here, we pre-train REGENT, a retrieval-augmented generalist agent that can adapt to unseen robotics and game-playing environments via in-context learning, without any finetuning. REGENT outperforms state-of-the-art generalist agents after pre-training on an order-of-magnitude fewer datapoints and with up to 3x fewer parameters. We also propose a strategy, inspired by adaptive control, to improve the robustness of the image encoder of REGENT, an essential component for handling environment variations. Finally, we bring REGENT to the real world by adding in-context adaptability to a pre-trained Vision Language Action model (VLA). Our method, called RICL, creates a RICL-VLA capable of adapting to completely new tasks with unseen objects and novel motions through retrieval-augmentation and in-context learning. Further task-specific RICL finetuning boosts performance, significantly improving over the base VLA directly finetuned on the same data. We conclude by outlining future directions to expand the envelope of tasks and environments to which a general AI agent can adapt.

Date of degree

2025

Date Range for Data Collection (Start Date)

Date Range for Data Collection (End Date)

Digital Object Identifier

Series name and number

Volume number

Issue number

Publisher

Publisher DOI

Journal Issues

Comments

Recommended citation