Goto

Collaborating Authors

 multi-lexsum


552ef803bef9368c29e53c167de34b55-Supplemental-Datasets_and_Benchmarks.pdf

Neural Information Processing Systems

For what purpose was the dataset created?Was therea specific task in mind? Was there aspecific gap that needed to be filled? Please provide a description.The Multi-LexSum dataset was curated to facilitate the development of automaticsummarization methods for civil rights lawsuits.Recent advances in document summarization have led to impressive results in generating ashort description for passages typically in hundreds of words. However, the source inputs forsummarizing civil right lawsuits are considerably longer: they can contain up to 70k words onaverage.


Multi-LexSum: Real-world Summaries of Civil Rights Lawsuits at Multiple Granularities

Neural Information Processing Systems

With the advent of large language models, methods for abstractive summarization have made great strides, creating potential for use in applications to aid knowledge workers processing unwieldy document collections. One such setting is the Civil Rights Litigation Clearinghouse (CRLC, https://clearinghouse.net),





Multi-LexSum: Real-world Summaries of Civil Rights Lawsuits at Multiple Granularities

Neural Information Processing Systems

With the advent of large language models, methods for abstractive summarization have made great strides, creating potential for use in applications to aid knowledge workers processing unwieldy document collections. One such setting is the Civil Rights Litigation Clearinghouse (CRLC, https://clearinghouse.net), Today, summarization in the CRLC requires extensive training of lawyers and law students who spend hours per case understanding multiple relevant documents in order to produce high-quality summaries of key events and outcomes. Motivated by this ongoing real-world summarization effort, we introduce Multi-LexSum, a collection of 9,280 expert-authored summaries drawn from ongoing CRLC writing. Multi-LexSum presents a challenging multi-document summarization task given the length of the source documents, often exceeding two hundred pages per case.


Multi-LexSum: Real-World Summaries of Civil Rights Lawsuits at Multiple Granularities

Shen, Zejiang, Lo, Kyle, Yu, Lauren, Dahlberg, Nathan, Schlanger, Margo, Downey, Doug

arXiv.org Artificial Intelligence

With the advent of large language models, methods for abstractive summarization have made great strides, creating potential for use in applications to aid knowledge workers processing unwieldy document collections. One such setting is the Civil Rights Litigation Clearinghouse (CRLC) (https://clearinghouse.net),which posts information about large-scale civil rights lawsuits, serving lawyers, scholars, and the general public. Today, summarization in the CRLC requires extensive training of lawyers and law students who spend hours per case understanding multiple relevant documents in order to produce high-quality summaries of key events and outcomes. Motivated by this ongoing real-world summarization effort, we introduce Multi-LexSum, a collection of 9,280 expert-authored summaries drawn from ongoing CRLC writing. Multi-LexSum presents a challenging multi-document summarization task given the length of the source documents, often exceeding two hundred pages per case. Furthermore, Multi-LexSum is distinct from other datasets in its multiple target summaries, each at a different granularity (ranging from one-sentence "extreme" summaries to multi-paragraph narrations of over five hundred words). We present extensive analysis demonstrating that despite the high-quality summaries in the training data (adhering to strict content and style guidelines), state-of-the-art summarization models perform poorly on this task. We release Multi-LexSum for further research in summarization methods as well as to facilitate development of applications to assist in the CRLC's mission at https://multilexsum.github.io.