Qwen1.5-MoE-A2.7B is an improved version of Qwen, a Large Language Model (LLM) series developed by the Qwen team at Alibaba Cloud. It performs on par with heavyweight 7B models like Mistral 7B and Qwen1.5-7B, despite having only 2.7 billion activated parameters. The architecture of Qwen1.5-MoE-A2.7B utilizes fine-grained experts and a generalized MoE routing paradigm.
•4m read time• From marktechpost.com
Sort: