AI Dynamics

Global AI News Aggregator

ProMoE: Scaling Diffusion Transformers with MoE for Visual Data

Why has scaling Diffusion Transformers with Mixture-of-Experts been so tricky for visual data? Researchers from Fudan University, Alibaba Group's Tongyi Lab, Zhejiang University, The University of Hong Kong, and MMLab just cracked the code! They introduce ProMoE, an MoE

→ View original post on X — @jiqizhixin,

Commentaires

Leave a Reply

Your email address will not be published. Required fields are marked *