How to run training jobs
You can either selectSupervised Fine-Tuning to train a model using labeled examples, or On-policy Distillation to train a student model using a teacher model for knowledge distillation.
Supervised fine-tuning (SFT)
To start an SFT job, initiate a training run from the Models page.- Click on
Train New Model. - In the Builder, select
Supervised Fine-Tuning. - Select the base model to fine-tune. Oumi offers a broad range of commonly used models.
- Choose your training dataset, and optionally select validation and test datasets. You can use uploaded datasets, synthesized data, or merged datasets.
- Select a training method. Oumi supports full fine-tuning (FFT) and parameter-efficient fine-tuning (PEFT), including LoRA.
- Adjust advanced hyperparameters (e.g.,
maximum steps,learning rate) if needed. - Review your configuration, (optionally) save it as a reusable recipe, and launch the training job.
On-policy distillation
To start an on-policy distillation job, initiate a training run from the Models page.- Click on
Train New Model. - In the Builder, select
On-Policy Distillation. - Leave
Training MethodonOn-Policy Distillation. - Choose your
Base ModelandTeacher Model. - Select your
Training Dataset. - Configure advanced settings (e.g.,
Training Settings,Distillation Settings,Parameter-Efficient Settings) if needed.
Please see On-Policy Distillation for more information regarding configuration optionss and settings.
Checking job status
After a training job launches, it will appear on the Activity log page with a status ofRunning.
When training completes, you can access your model from the Model page.