LongLLMLingua: Accelerating and Enhancing LLMs in Long Context Scenarios via Prompt Compression
Original Paper: https://arxiv.org/abs/2310.06839 By: Huiqiang Jiang, Qianhui Wu, Xufang Luo, Dongsheng Li, Chin-Yew Lin, Yuqing Yang, Lili Qiu Abstract: In long context scenarios, large language models (LLMs) face three main challenges: higher computational/financial cost, longer latency, and inferior performance. Some studies reveal that the