Researchers from Kunlun Inc. have introduced DiT-MoE, an advanced version of the DiT architecture for image generation, that incorporates sparse Mixture of Experts (MoE) layers to enhance efficiency and performance. This new model significantly outperforms previous architectures in conditional image generation tasks, using
Sort: