](https://deep-paper.org/en/paper/643_from_weight_based_to_state-1884/images/cover.png)
Beyond LoRA: How State-Based Control Unlocks Training 8B Models on Consumer GPUs
Introduction If you have ever tried to fine-tune a Large Language Model (LLM) on your local machine, you have likely run into the dreaded “CUDA Out of Memory” error. Modern models like LLaMA-3 are incredibly capable, but they are also massive. Even with the advent of Parameter-Efficient Fine-Tuning (PEFT) methods like Low-Rank Adaptation (LoRA), the memory requirements often exceed what is available on standard consumer-grade hardware (like an NVIDIA RTX 3090 or 4090). ...
](https://deep-paper.org/en/paper/626_mixture_of_lookup_experts-1883/images/cover.png)
](https://deep-paper.org/en/paper/15148_accelerating_llm_inferen-1881/images/cover.png)
](https://deep-paper.org/en/paper/7375_outlier_gradient_analysis-1880/images/cover.png)
](https://deep-paper.org/en/paper/2325_can_mllms_reason_in_multi-1879/images/cover.png)
](https://deep-paper.org/en/paper/2505.09010/images/cover.png)
](https://deep-paper.org/en/paper/2505.23760/images/cover.png)
](https://deep-paper.org/en/paper/2503.06366/images/cover.png)
](https://deep-paper.org/en/paper/3607_videorope_what_makes_for_-1874/images/cover.png)
](https://deep-paper.org/en/paper/2508.08252/images/cover.png)
](https://deep-paper.org/en/paper/2503.07067/images/cover.png)
](https://deep-paper.org/en/paper/2405.15991/images/cover.png)
](https://deep-paper.org/en/paper/2407.00397/images/cover.png)
](https://deep-paper.org/en/paper/2502.09376/images/cover.png)
](https://deep-paper.org/en/paper/6368_an_improved_clique_pickin-1865/images/cover.png)
](https://deep-paper.org/en/paper/2506.17204/images/cover.png)
](https://deep-paper.org/en/paper/2502.06737/images/cover.png)
](https://deep-paper.org/en/paper/2502.08532/images/cover.png)
](https://deep-paper.org/en/paper/2506.05503/images/cover.png)
](https://deep-paper.org/en/paper/2502.05352/images/cover.png)