Incidental Supervision for Natural Language Understanding
Degree type
Graduate group
Discipline
Subject
Funder
Grant number
License
Copyright date
Distributor
Related resources
Author
Contributor
Abstract
Acquiring human annotations for natural language understanding (NLU) tasks is often labor-intensive and demands significant domain-specific expertise, making it crucial to obtain supervision from indirect signals to improve target task performance. This dissertation presents a novel approach for enhancing NLU by harnessing on the power of incidental supervision signals, which are present in the data and the environment, regardless of the specific tasks being considered. The primary aim of this research is to deepen our understanding of incidental supervision signals and to develop efficient algorithms for their acquisition, selection, and usage in NLU tasks. This problem presents numerous challenges, including the intricate nature of natural language and the inherent disparities between incidental supervision signals and target tasks. To tackle these challenges, this dissertation employs a multifaceted approach. First, we demonstrate the feasibility of utilizing cost-effective signals to enhance various target tasks. Specifically, we retrieve signals from sentence-level question-answer pairs to help NLU tasks via two types of sentence encoding approaches, depending on whether the target task involves single- or multi-sentence input. Second, we introduce a unified informativeness measure to quantitatively assess the effectiveness of diverse incidental supervision signals for a given target task. This approach offers a promising way to determine, ahead of learning, which supervision signals would be beneficial. Finally, we present a suite of efficient algorithms for exploiting distinct types of incidental supervision signals, including a weighting strategy to enhance sample efficiency in cross-task learning and a post-processing technique for faithful large language model inference with external knowledge.