](https://deep-paper.org/en/paper/2407.17817/images/cover.png)
Why Your LLM Can't Keep a Secret: The Science of Verbatim Memorization
In the world of Large Language Models (LLMs), there is a ghost in the machine. Sometimes, models like GPT-4 or Claude don’t just generate novel text—they recite specific training data word-for-word. This phenomenon, known as verbatim memorization, ranges from the innocuous (reciting the Gettysburg Address) to the legally hazardous (reproducing copyrighted code or private identifying information). For years, researchers have treated this as a bug to be squashed. The prevailing assumption has been that specific “bad” weights or neurons are hoarding these memories, and if we could just locate and prune them, the problem would vanish. ...
](https://deep-paper.org/en/paper/2402.04401/images/cover.png)
](https://deep-paper.org/en/paper/file-2939/images/cover.png)
](https://deep-paper.org/en/paper/2410.02499/images/cover.png)
](https://deep-paper.org/en/paper/2402.13148/images/cover.png)
](https://deep-paper.org/en/paper/2410.05639/images/cover.png)
](https://deep-paper.org/en/paper/2407.07840/images/cover.png)
](https://deep-paper.org/en/paper/2311.09630/images/cover.png)
](https://deep-paper.org/en/paper/file-2933/images/cover.png)
](https://deep-paper.org/en/paper/2410.05162/images/cover.png)
](https://deep-paper.org/en/paper/file-2931/images/cover.png)
](https://deep-paper.org/en/paper/file-2930/images/cover.png)
](https://deep-paper.org/en/paper/file-2929/images/cover.png)
](https://deep-paper.org/en/paper/2410.17859/images/cover.png)
](https://deep-paper.org/en/paper/2408.05346/images/cover.png)
](https://deep-paper.org/en/paper/2407.06380/images/cover.png)
](https://deep-paper.org/en/paper/2406.13236/images/cover.png)
](https://deep-paper.org/en/paper/2410.05269/images/cover.png)
](https://deep-paper.org/en/paper/2407.21417/images/cover.png)
](https://deep-paper.org/en/paper/file-2922/images/cover.png)