NVIDIA VP of Generative AI Kari Briski explains why a chip maker builds LLMs: the hardware-software co-design feedback loop requires deeply understanding workloads to optimize them. She covers NVIDIA's Nemotron model family (Nano, Super, Ultra), the benefits of training at reduced floating-point precision (NVFP4) over

27m read time From stackoverflow.blog
Post cover image
Table of contents
TRANSCRIPT

Sort: