University of Cambridge > Talks.cam > Natural Language Processing Reading Group > NLP Reading Group: Large-margin Learning of Submodular Summarization methods

NLP Reading Group: Large-margin Learning of Submodular Summarization methods

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact Jimme Jardine.

This week Helen will be talking about :

Large-margin Learning of Submodular Summarization methods http://arxiv.org/abs/1110.2162

In this paper, we present a supervised learning approach to training submodular scoring functions for extractive multi-document summarization. By taking a structured predicition approach, we provide a large-margin method that directly optimizes a convex relaxation of the desired performance measure. The learning method applies to all submodular summarization methods, and we demonstrate its effectiveness for both pairwise as well as coverage-based scoring functions on multiple datasets. Compared to state-of-the-art functions that were tuned manually, our method significantly improves performance and enables high-fidelity models with numbers of parameters well beyond what could reasonbly be tuned by hand.

This talk is part of the Natural Language Processing Reading Group series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2024 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity