Keyboard shortcuts

Press or to navigate between chapters

Press S or / to search in the book

Press ? to show this help

Press Esc to hide this help

Model Merging

After training, the LoRA adapters are saved as a separate file. The run_model_merge function in train_pipeline.py performs the final step: it loads the base model in its original precision (bfloat16) and merges the trained adapter weights into it. The result is a single, standalone fine-tuned model that can be deployed without requiring the PEFT library.