Benchmark Deep Dive: Telegraph English vs LLMLingua-2
TL;DR: Telegraph English outperforms LLMLingua-2 across all models and datasets, losing only 1-3pp on headline facts and 3-11pp on fine details—still ahead of LLML2 at every compression level. TE's symbolic structure (→, caps, atomic facts) keeps logical chains and whole facts intact, helping weaker models avoid hallucinations. LLML2's token deletion can break chains, forcing smaller models to "fill in the blanks." The gap widens to up to 11pp on smaller models and detail-heavy tasks.