](https://deep-paper.org/en/paper/2407.02987/images/cover.png)
LoRA-Guard: Achieving On-Device AI Safety with Parameter-Efficient Adaptation
Introduction The rapid evolution of Large Language Models (LLMs) has brought us capable conversational assistants, coding partners, and creative writers. However, this capability comes with a significant caveat: without careful alignment, these models can generate toxic, offensive, or illegal content. While “safety tuning” (like Reinforcement Learning from Human Feedback) helps, it isn’t a silver bullet. Jailbreaks—cleverly crafted prompts designed to bypass safety filters—remain a persistent threat. To combat this, the industry has turned to guardrails: separate, dedicated models that monitor the conversation and flag harmful content. The problem? Running a massive LLM is already computationally expensive. Running a second massive model just to police the first one is often impossible, especially on resource-constrained devices like mobile phones or laptops. ...
](https://deep-paper.org/en/paper/2407.18940/images/cover.png)
](https://deep-paper.org/en/paper/2410.18385/images/cover.png)
](https://deep-paper.org/en/paper/2406.08818/images/cover.png)
](https://deep-paper.org/en/paper/file-3314/images/cover.png)
](https://deep-paper.org/en/paper/2405.03279/images/cover.png)
](https://deep-paper.org/en/paper/2410.08905/images/cover.png)
](https://deep-paper.org/en/paper/2406.13560/images/cover.png)
](https://deep-paper.org/en/paper/file-3310/images/cover.png)
](https://deep-paper.org/en/paper/2401.07103/images/cover.png)
](https://deep-paper.org/en/paper/2409.16198/images/cover.png)
](https://deep-paper.org/en/paper/file-3307/images/cover.png)
](https://deep-paper.org/en/paper/file-3306/images/cover.png)
](https://deep-paper.org/en/paper/file-3305/images/cover.png)
](https://deep-paper.org/en/paper/file-3304/images/cover.png)
](https://deep-paper.org/en/paper/2407.01906/images/cover.png)
](https://deep-paper.org/en/paper/2307.00279/images/cover.png)
](https://deep-paper.org/en/paper/2410.15148/images/cover.png)
](https://deep-paper.org/en/paper/2406.17419/images/cover.png)
](https://deep-paper.org/en/paper/2411.03550/images/cover.png)