CLIPPER: Compression enables long-context synthetic data generation
Abstract
LLM developers are increasingly reliant on synthetic data, but generating high-quality data for complex long-context reasoning tasks remains challenging. We introduce CLIPPER, a compression-based approach for generating synthetic data tailored to narrative claim verification - a task that requires reasoning over a book to verify a given claim. Instead of generating claims directly from the raw text of the book, which results in artifact-riddled claims, CLIPPER first compresses the book into chapter outlines and book summaries and then uses these intermediate representations to generate complex claims and corresponding chain-of-thoughts. Compared to naive approaches, CLIPPER produces claims that are more valid, grounded, and complex. Using CLIPPER, we construct a dataset of 19K synthetic book claims paired with their source texts and chain-of-thought reasoning, and use it to fine-tune three open-weight models. Our best model achieves breakthrough results on narrative claim verification (from 28% to 76% accuracy on our test set) and sets a new state-of-the-art for sub-10B models on the NoCha leaderboard. Further analysis shows that our models generate more detailed and grounded chain-of-thought reasoning while also improving performance on other narrative understanding tasks (e.g., NarrativeQA).
Community
CLIPPER is an approach to generating instruction-following data by compressing long-form documents (e.g., books) into smaller, information-rich representations (e.g. chapter outlines), which are then used to create grounded instructions for tasks like narrative claim verification. Open-source models fine-tuned on CLIPPER data show substantial gains in verification and narrative understanding: Our best model that is fine-tuned on LLaMA-3.1-8B-Instruct sets the new state-of-the-art for sub-10B models on NoCha, a long-form narrative claim verification benchmark.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- LongFaith: Enhancing Long-Context Reasoning in LLMs with Faithful Synthetic Data (2025)
- BARE: Combining Base and Instruction-Tuned Language Models for Better Synthetic Data Generation (2025)
- FactCG: Enhancing Fact Checkers with Graph-Based Multi-Hop Data (2025)
- Chain-of-Retrieval Augmented Generation (2025)
- Emulating Retrieval Augmented Generation via Prompt Engineering for Enhanced Long Context Comprehension in LLMs (2025)
- Rationalization Models for Text-to-SQL (2025)
- Domaino1s: Guiding LLM Reasoning for Explainable Answers in High-Stakes Domains (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 4
Datasets citing this paper 2
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper