](https://deep-paper.org/en/paper/2508.15487/images/cover.png)
Beyond Left-to-Right: Introducing Dream 7B, a Powerful New Diffusion LLM
For years, large language models (LLMs) have relied on a single fundamental idea: autoregression. Models such as GPT-4, LLaMA, and Qwen generate text one word at a time, moving from left to right—much like how a person might write a sentence. This approach has driven remarkable progress, but it also carries inherent limitations. When a model can only see the past, it struggles with tasks requiring global consistency, long-term planning, or satisfying complex constraints. ...