Automatic Evaluation of Linguistic Quality in Multi-Document Summarization

Loading...
Thumbnail Image
Penn collection
Departmental Papers (CIS)
Degree type
Discipline
Subject
Computer Sciences
Funder
Grant number
License
Copyright date
Distributor
Related resources
Author
Pitler, Emily
Louis, Annie
Contributor
Abstract

To date, few attempts have been made to develop and validate methods for automatic evaluation of linguistic quality in text summarization. We present the first systematic assessment of several diverse classes of metrics designed to capture various aspects of well-written text. We train and test linguistic quality models on consecutive years of NIST evaluation data in order to show the generality of results. For grammaticality, the best results come from a set of syntactic features. Focus, coherence and referential clarity are best evaluated by a class of features measuring local coherence on the basis of cosine similarity between sentences, coreference information, and summarization specific features. Our best results are 90% accuracy for pairwise comparisons of competing systems over a test set of several inputs and 70% for ranking summaries of a specific input.

Advisor
Date of presentation
2010-07-01
Conference name
Departmental Papers (CIS)
Conference dates
2023-05-17T07:16:50.000
Conference location
Date Range for Data Collection (Start Date)
Date Range for Data Collection (End Date)
Digital Object Identifier
Series name and number
Volume number
Issue number
Publisher
Publisher DOI
Journal Issue
Comments
Pitler, E., Louis, A., & Nenkova, A., Automatic Evaluation of Linguistic Quality in Multi-Document Summarization, 48th Annual Meeting of the Association for Computational Linguistics, July 2010, doi: http://www.aclweb.org/anthology/P10-1056
Recommended citation
Collection