NVIDIA has published enterprise reference architectures (RAs) for building on-premises AI factories. Three configurations are outlined: RTX PRO AI Factory for small-to-medium workloads (up to 256 GPUs, air-cooled), HGX AI Factory for large-scale LLM training and inference (Blackwell Ultra GPUs, 800 Gb/s networking), and NVL72 AI Factory for exascale workloads (72 Blackwell Ultra GPUs per rack, liquid-cooled, unified NVLink fabric). These RAs provide end-to-end guidance covering compute, networking, storage, and software integration, and are validated by system partners through NVIDIA's Design Review Board. The goal is to reduce deployment timelines, lower TCO, and help enterprises move from proof-of-concept to production AI operations.

7m read timeFrom developer.nvidia.com
Post cover image
Table of contents
Enterprise RAs form the foundation of AI factoriesNVIDIA HGX AI Factory: Breakthrough performance for enterprise AINVIDIA NVL72 AI Factory: Powering exascale AIFaster deployment and lower TCOsReady to get started?

Sort: