Fine-tuning
The process of taking a pre-trained AI model and adapting it to a specific task or dataset. Fine-tuning requires significantly more VRAM and compute than inference because the GPU must store model weights, gradients, and optimizer states simultaneously. For consumer hardware, 24 GB VRAM is the practical minimum for fine-tuning 7B-parameter models; larger models need multi-GPU setups or cloud compute.