COMPOSITIONAL GENERALIZATION IN INSTRUCTION FOLLOWING TASKS
Degree type
Graduate group
Discipline
Subject
Machine Learning
Natural Language Processing
Funder
Grant number
License
Copyright date
Distributor
Related resources
Author
Contributor
Abstract
Understanding instructions expressed in natural language is a fundamental task in artificialintelligence. A key feature of natural language that humans use while giving and following instructions is compositionality: the capacity to understand and produce a potentially infinite number of novel combinations from familiar components. This ability is instrumental in being able to learn from limited data and is crucial for instruction following robots to function in the real-world. This dissertation studies the compositional generalization abilities of machine learning models in various instruction following tasks. We study the various dimensions of compositionality for a diverse set of instruction follow- ing tasks of varying complexity: semantic parsing in synthetic languages, natural language instruction following in blocks world and vision-and-language navigation in complex indoor environments. We demonstrate empirically that existing systems for these tasks, while performant on the standard iid test set requiring mere interpolation, do not compositionally generalize, which requires extrapolation. We then present different strategies to induce compositionality, ranging from data augmentation, to auxiliary tasks, to a simple neuro-symbolic algorithm. We present a compositional spatial representation language and discuss how using such a rich symbolic representation as auxiliary supervision can help generalization in complex, real-world, multi-modal instruction following tasks. Finally, we aim to develop a more foundational understanding of robust generalization by focusing on the task of learning regular languages, where we study the benefits of compositional models over end-to-end ones, from both theoretical and empirical perspectives.