Adapting LLMs for Efficient Context Processing through Soft Prompt Compression

Open in new window