LoRA Adapter
The small set of trained weight matrices produced by LoRA fine-tuning. A LoRA adapter is typically just 10–100 MB compared to the multi-gigabyte base model, and can be swapped in and out at inference time to customize model behavior for different tasks. You can collect multiple adapters for different use cases and apply them to the same base model, making it a flexible system for local AI customization.