Technique for Parameter Efficient Fine Tuning/PEFT, specifically for Large Language Model/Diffusion Models such that they can be adapted to novel tasks.

Freezes base model weights, and then trains new, tiny, adapter matrices. This reduces memory requirements and training time significantly.

You are also left with a very small LoRa adapter file, which for a given base model, can be switched out quite easily and quickly.