NVIDIA introduces Nemotron 3 Nano 4B, a 4-billion-parameter hybrid Mamba-Transformer language model optimized for edge deployment on NVIDIA Jetson, GeForce RTX, and DGX Spark platforms. The model was compressed from the 9B Nemotron Nano v2 using the Nemotron Elastic structured pruning and distillation framework, which jointly

8m read timeFrom huggingface.co
Post cover image
Table of contents
Training Recipe for Nemotron 3 Nano 4BBoosting Efficiency with QuantizationTry It Now!

Sort: