](https://deep-paper.org/en/paper/2412.15214/images/cover.png)
Beyond Flat Video: Mastering 3D Trajectory Control in Generative AI with LeviTor
Introduction In the rapidly evolving world of Generative AI, we have moved quickly from creating static images to generating full-motion video. Tools like Sora, Runway, and Stable Video Diffusion have shown us that AI can dream up dynamic scenes. However, for these tools to be useful in professional workflows—like filmmaking, game design, or VR—random generation isn’t enough. We need control. Specifically, we need to tell the AI exactly where an object should move. This concept, known as “drag-based interaction,” allows users to click a point on an image and drag it to a new location, signaling the model to animate that movement. ...
](https://deep-paper.org/en/paper/2409.16434/images/cover.png)
](https://deep-paper.org/en/paper/2412.00054/images/cover.png)
](https://deep-paper.org/en/paper/file-2108/images/cover.png)
](https://deep-paper.org/en/paper/2504.02697/images/cover.png)
](https://deep-paper.org/en/paper/file-2106/images/cover.png)
](https://deep-paper.org/en/paper/2503.21099/images/cover.png)
](https://deep-paper.org/en/paper/2503.15019/images/cover.png)
](https://deep-paper.org/en/paper/2412.20651/images/cover.png)
](https://deep-paper.org/en/paper/file-2102/images/cover.png)
](https://deep-paper.org/en/paper/2503.06956/images/cover.png)
](https://deep-paper.org/en/paper/file-2100/images/cover.png)
](https://deep-paper.org/en/paper/2501.18954/images/cover.png)
](https://deep-paper.org/en/paper/2503.21076/images/cover.png)
](https://deep-paper.org/en/paper/file-2097/images/cover.png)
](https://deep-paper.org/en/paper/2411.14205/images/cover.png)
](https://deep-paper.org/en/paper/2411.16198/images/cover.png)
](https://deep-paper.org/en/paper/2505.24315/images/cover.png)
](https://deep-paper.org/en/paper/2502.20390/images/cover.png)
](https://deep-paper.org/en/paper/2412.12087/images/cover.png)