This post provides a deep dive into fine-tuning Llama-3 with ORPO. It covers the code breakdown, customization, notebook login, defining tasks and preferences, loading tokenizer and model, preparing ORPO trainer, training the model, and merging the QLoRA adapter with the base model. The post also mentions additional considerations such as evaluation and hardware requirements.

6m read time From blog.gopenai.com
Post cover image

Sort: