](https://deep-paper.org/en/paper/2407.09014/images/cover.png)
Squeezing the Truth: How COMPACT Makes RAG Smarter and Faster
Introduction In the rapidly evolving world of Large Language Models (LLMs), Retrieval-Augmented Generation (RAG) has become the gold standard for grounding AI responses in reality. By fetching relevant documents from an external database, we can prevent hallucinations and give models access to up-to-date information. However, there is a catching point: the “context window.” While modern models boast about handling 100k or even 1 million tokens, filling that context comes with significant downsides. It is expensive, increases latency, and paradoxically, often confuses the model. Known as the “Lost in the Middle” phenomenon, LLMs struggle to find specific needles in massive haystacks of retrieved text. ...
](https://deep-paper.org/en/paper/2406.12074/images/cover.png)
](https://deep-paper.org/en/paper/2410.05851/images/cover.png)
](https://deep-paper.org/en/paper/2410.23844/images/cover.png)
](https://deep-paper.org/en/paper/2410.03077/images/cover.png)
](https://deep-paper.org/en/paper/2402.15002/images/cover.png)
](https://deep-paper.org/en/paper/2410.02428/images/cover.png)
](https://deep-paper.org/en/paper/2407.01300/images/cover.png)
](https://deep-paper.org/en/paper/2409.19715/images/cover.png)
](https://deep-paper.org/en/paper/2410.02184/images/cover.png)
](https://deep-paper.org/en/paper/2402.02172/images/cover.png)
](https://deep-paper.org/en/paper/2401.10065/images/cover.png)
](https://deep-paper.org/en/paper/2409.19753/images/cover.png)
](https://deep-paper.org/en/paper/2406.17626/images/cover.png)
](https://deep-paper.org/en/paper/2408.15992/images/cover.png)
](https://deep-paper.org/en/paper/2406.07054/images/cover.png)
](https://deep-paper.org/en/paper/2403.13583/images/cover.png)
](https://deep-paper.org/en/paper/2410.03457/images/cover.png)
](https://deep-paper.org/en/paper/2410.06741/images/cover.png)
](https://deep-paper.org/en/paper/2411.01176/images/cover.png)