LLavaCode: Compressed Code Representations for Retrieval-Augmented Code Generation
Cherniuk, Daria, Sukhorukov, Nikita, Sushko, Nikita, Gusak, Daniil, Sivtsov, Danil, Tutubalina, Elena, Frolov, Evgeny
–arXiv.org Artificial Intelligence
Retrieval-augmented generation has emerged as one of the most effective approaches for code completion, particularly when context from a surrounding repository is essential. However, incorporating context significantly extends sequence length, leading to slower inference - a critical limitation for interactive settings such as IDEs. In this work, we introduce LlavaCode, a framework that compresses code into compact, semantically rich representations interpretable by code LLM, enhancing generation quality while reducing the retrieved context to only a few compressed single-token vectors. Using a small projector module we can significantly increase the EM and ES metrics of coding model with negligible latency increase. Our experiments demonstrate that compressed context enables 20-38% reduction in Time-to-First-Token (TTFT) on line completion tasks compared to full-RAG pipelines.
arXiv.org Artificial Intelligence
Oct-23-2025
- Country:
- Asia > Russia (0.05)
- Europe
- Russia > Central Federal District
- Moscow Oblast > Moscow (0.05)
- Ukraine > Kyiv Oblast
- Kyiv (0.04)
- Russia > Central Federal District
- North America > United States
- Massachusetts > Middlesex County > Cambridge (0.04)
- Genre:
- Research Report (0.82)
- Technology: