Llama 2, an assortment of pretrained and fine-tuned large language models (LLMs) with sizes varying from 7 billion to 70 billion parameters, are specifically designed for dialogue applications. These models surpass the performance of existing open-source chat models on most benchmarks, and could potentially replace closed-source models.
Table of contents
Supervised Fine-Tuning (SFT)Human Preference Data CollectionReward ModelingReward Model ResultsSort: