Overview
A training recipe captures a complete model training configuration (base model, datasets, training method, and hyperparameters) as a reusable JSON template. Running a recipe produces a consistent, reproducible training job without requiring you to reconfigure inputs each time. Training recipes are saved from the Oumi Builder and accessed from the Recipes page. For details on saving and running a recipe from the UI, see Training Recipes. For the full schema reference, see Model Training Recipe Schema.Common recipe patterns
Supervised fine-tuning with LoRA (PEFT)
The most common starting point. LoRA reduces GPU memory requirements and trains faster than full fine-tuning, making it ideal for rapid iteration.Full-weight fine-tuning (FFT)
Updates all model parameters. Use when you need deep behavioral changes or have sufficient compute available.On-policy distillation
Trains a smaller student model guided by a stronger teacher. RequirestrainerType: "opd". See On-Policy Distillation for configuration details.
Tips
- Start with LoRA: switch to FFT only if LoRA quality is insufficient for your task.
- Use
evalStrategy: "epoch"for small datasets; use"steps"with a reasonableevalStepsfor large ones. - Set
runName: descriptive names make it easier to compare runs in the activity log. - Use
inferenceSeedfor reproducible results across recipe runs. - Enable
enableGradientCheckpointingwhen GPU memory is constrained; it slows training slightly but allows larger batch sizes.