](https://deep-paper.org/en/paper/2403.00824/images/cover.png)
Mapping the Mind of an LLM: How Information Flow Routes Reveal Model Inner Workings
The inner workings of Large Language Models (LLMs) often feel like a black box. We feed a prompt into one end, and a coherent response magically appears at the other. We know the architecture—Transformers, attention heads, feed-forward networks—but understanding exactly how a specific input token influences a specific output prediction remains one of the hardest challenges in AI research. Traditionally, researchers have tried to reverse-engineer these models using “circuits”—subgraphs of the model responsible for specific tasks. However, finding these circuits is usually a manual, labor-intensive process that requires human intuition to design specific test cases. ...
](https://deep-paper.org/en/paper/2410.01518/images/cover.png)
](https://deep-paper.org/en/paper/file-3203/images/cover.png)
](https://deep-paper.org/en/paper/2401.11206/images/cover.png)
](https://deep-paper.org/en/paper/2404.11095/images/cover.png)
](https://deep-paper.org/en/paper/file-3200/images/cover.png)
](https://deep-paper.org/en/paper/2404.10877/images/cover.png)
](https://deep-paper.org/en/paper/2503.16043/images/cover.png)
](https://deep-paper.org/en/paper/2405.10512/images/cover.png)
](https://deep-paper.org/en/paper/file-3196/images/cover.png)
](https://deep-paper.org/en/paper/2311.07237/images/cover.png)
](https://deep-paper.org/en/paper/2410.00025/images/cover.png)
](https://deep-paper.org/en/paper/file-3193/images/cover.png)
](https://deep-paper.org/en/paper/2407.15343/images/cover.png)
](https://deep-paper.org/en/paper/file-3191/images/cover.png)
](https://deep-paper.org/en/paper/file-3190/images/cover.png)
](https://deep-paper.org/en/paper/2405.19842/images/cover.png)
](https://deep-paper.org/en/paper/2410.15801/images/cover.png)
](https://deep-paper.org/en/paper/2410.09318/images/cover.png)
](https://deep-paper.org/en/paper/2407.02814/images/cover.png)