](https://deep-paper.org/en/paper/2409.04206/images/cover.png)
Fast Forward: How to Speed Up LLM Finetuning by Just Keeping Going
Training Large Language Models (LLMs) is computationally expensive. Even as we’ve moved from training from scratch to fine-tuning pre-trained models, the cost in terms of time and GPU compute (FLOPs) remains a massive barrier for students and researchers. To mitigate this, the community adopted Parameter Efficient Fine-Tuning (PEFT) methods, with LoRA (Low-Rank Adaptation) being the undisputed champion. LoRA reduces the memory footprint significantly by freezing the main model weights and training only a small subset of parameters. But here is the catch: while LoRA saves memory, it doesn’t necessarily speed up the training process itself by a huge margin. You still have to run thousands of iterations of Stochastic Gradient Descent (SGD). ...
](https://deep-paper.org/en/paper/2406.11370/images/cover.png)
](https://deep-paper.org/en/paper/file-3086/images/cover.png)
](https://deep-paper.org/en/paper/2402.02420/images/cover.png)
](https://deep-paper.org/en/paper/file-3084/images/cover.png)
](https://deep-paper.org/en/paper/file-3083/images/cover.png)
](https://deep-paper.org/en/paper/2209.00840/images/cover.png)
](https://deep-paper.org/en/paper/2308.04265/images/cover.png)
](https://deep-paper.org/en/paper/2404.11184/images/cover.png)
](https://deep-paper.org/en/paper/2408.12168/images/cover.png)
](https://deep-paper.org/en/paper/2406.15657/images/cover.png)
](https://deep-paper.org/en/paper/file-3077/images/cover.png)
](https://deep-paper.org/en/paper/2408.10276/images/cover.png)
](https://deep-paper.org/en/paper/2410.10859/images/cover.png)
](https://deep-paper.org/en/paper/file-3074/images/cover.png)
](https://deep-paper.org/en/paper/file-3073/images/cover.png)
](https://deep-paper.org/en/paper/file-3072/images/cover.png)
](https://deep-paper.org/en/paper/2405.15012/images/cover.png)
](https://deep-paper.org/en/paper/2404.03868/images/cover.png)
](https://deep-paper.org/en/paper/file-3069/images/cover.png)