Papers
Topics
Authors
Recent
Search
2000 character limit reached

Learning Mixtures of Submodular Shells with Application to Document Summarization

Published 16 Oct 2012 in cs.LG, cs.CL, cs.IR, and stat.ML | (1210.4871v1)

Abstract: We introduce a method to learn a mixture of submodular "shells" in a large-margin setting. A submodular shell is an abstract submodular function that can be instantiated with a ground set and a set of parameters to produce a submodular function. A mixture of such shells can then also be so instantiated to produce a more complex submodular function. What our algorithm learns are the mixture weights over such shells. We provide a risk bound guarantee when learning in a large-margin structured-prediction setting using a projected subgradient method when only approximate submodular optimization is possible (such as with submodular function maximization). We apply this method to the problem of multi-document summarization and produce the best results reported so far on the widely used NIST DUC-05 through DUC-07 document summarization corpora.

Citations (187)

Summary

  • The paper proposes a novel method for learning mixtures of submodular shells within a large-margin structured-prediction framework.
  • It employs a projected subgradient method to optimize mixture weights and provides risk-bound guarantees.
  • The approach sets new benchmarks in multi-document summarization (NIST DUC datasets), significantly outperforming previous methods in ROUGE scores.

Overview of Learning Mixtures of Submodular Shells with Application to Document Summarization

This paper presents a nuanced approach to learning mixtures of submodular functions, specifically submodular shells, and applies this framework to document summarization tasks. Submodular functions possess a diminishing returns property, which makes them particularly suitable for modeling tasks that require efficient selection of information subsets under constraints. The method proposed by Lin and Bilmes builds upon previous research by extending the concept of submodular mixtures to submodular shells, offering a structured way to optimize these functions using data-driven techniques.

Methodology

The crux of the paper lies in the development of an algorithm capable of learning mixture weights over submodular shells within a large-margin structured-prediction setting. Each submodular shell is a parametric function that can be instantiated with a specific ground set and parameters, thereby inducing a submodular function. The authors employ a projected subgradient method to optimize the mixture weights while providing risk-bound guarantees when exact submodular optimization is intractable.

Numerical Results

Strong numerical evidence is offered to demonstrate the effectiveness of the approach. The method was applied to multi-document summarization tasks using datasets from NIST DUC (2005-2007). The results indicate that the learned submodular shell mixtures outperform existing methods, setting new benchmarks in ROUGE score recall and F-measure metrics across multiple DUC datasets. Specifically, for DUC-05 query-focused summarization, the approach achieved a ROUGE-2 recall of 8.44%, significantly surpassing the previous best of 7.82%.

Implications and Future Directions

The application of submodular shell mixtures to document summarization reveals the potential of these functions in handling complex structured prediction problems. This has implications for improving extractive summarization approaches and potentially extending to other domains where subset selection under constraints is pivotal, such as sensor placement or social network analysis.

From a theoretical perspective, this work aids in bridging the gap between submodular function theory and practical applications in machine learning. The paper suggests that future research could explore the use of submodular shell mixtures in more diverse settings, potentially incorporating dynamic ground sets or complex objectives that require intricate submodular formulations.

Conclusion

Lin and Bilmes have contributed substantially to the field by proposing a robust framework for learning submodular shell mixtures, refining structured inferred learning tasks with submodular properties. This paper's approach and results underscore the adaptability and expressiveness of submodular shell mixtures in optimizing document summarization, providing a platform for future explorations and advancements in structured machine learning applications.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Authors (2)

Collections

Sign up for free to add this paper to one or more collections.