config/
directorydefault_config.json
or config_online_qwq_phi4_uld.json
)"logits_file"
, you must first generate the teacher logits by following the steps in the Generating Teacher Logits section.distillation-volume
). The corresponding Modal scripts (scripts/modal/distill_logits.py
or scripts/modal/generate_logits.py
) must be configured to mount this volume and access the files from the volume path.distill_logits.py
script with your configuration:
training.output_dir
with a subdirectory called final-distilled-checkpoint
. You can load it like any Hugging Face model:
forward_kl
(fkl
) loss type for distillation when using pre-computed logits.