SmolLM2 is a family of compact language models ranging from 135M to 1.7B parameters, designed for on-device use with versatile capabilities. The SmolLM2-1.7B-Instruct model can be used as an assistant via various tools and frameworks. Detailed instructions for pre-training, fine-tuning, and using these models are provided.

3m read timeFrom github.com
Post cover image
Table of contents
Table of ContentsUsageSmol-toolsPre-trainingFine-tuningEvaluationSynthetic data pipelines

Sort: