Gradual Forgetting: Logarithmic Compression for Extending Transformer Context Windows
–arXiv.org Artificial Intelligence
Most approaches to long-context processing increase the complexity of the transformer's internal architecture by integrating mechanisms such as recurrence or auxiliary memory modules. In this work, we introduce an alternative approach that modifies the input representation itself, rather than the transformer architecture. Inspired by cognitive models of human memory, our method applies a scale-invariant logarithmic compression to the input tokens. The resulting compressed representation is processed by a standard, unmodified transformer, preserving architectural simplicity. We evaluate this approach on the WikiText-103 and PG-19 language modeling benchmarks, showing a reduction in perplexity compared to uncompressed baselines. Moreover, performance improves consistently with longer compressed temporal contexts, showing that input-level logarithmic compression is a simple and effective way to extend a transformer's long-range memory.
arXiv.org Artificial Intelligence
Oct-28-2025
- Country:
- North America > United States
- Indiana (0.04)
- South America > Chile
- North America > United States
- Genre:
- Research Report (0.50)
- Industry:
- Health & Medicine > Therapeutic Area > Neurology (0.69)
- Technology: