China's open-source AI ecosystem has shifted toward Mixture-of-Experts (MoE) architectures as the default choice, prioritizing cost-performance balance over maximum capability. Leading organizations expanded beyond text models into multimodal domains (video, audio, 3D), with growing emphasis on small models (0.5B-30B parameters) for practical deployment. Apache 2.0 became the standard license, reducing friction for production use. A significant strategic shift emerged toward hardware-first development, with models increasingly optimized for domestic Chinese chips (Huawei Ascend, Cambricon, Baidu Kunlun) in both inference and training. Companies are open-sourcing production-grade serving systems and infrastructure, moving competition from isolated model performance to full-stack ecosystem design.
Table of contents
Mixture of Experts (MoE) as the Default ChoiceThe Rush for Supremacy by ModalityBig Preferences for Small ModelsMore Permissive Open Source LicensesFrom Model-First to Hardware-FirstReconstruction In Progress1 Comment
Sort: