GLM-4.7-Flash is a 30B-A3B MoE (Mixture of Experts) language model that balances performance and efficiency for lightweight deployment. It demonstrates strong benchmark performance across tasks like AIME 25, GPQA, and SWE-bench Verified, outperforming comparable models in several categories. The model supports local deployment
Sort: