
Alibaba Releases Open-Source Wan 2.2 Video Generator Using MoE Architecture
Alibaba’s research unit has released Wan 2.2, an open-source large-scale video generation model distributed under the Apache 2.0 license. The system is built on a Mixture-of-Experts architecture that expands model capacity without increasing compute requirements, making it the first publicly available video generator to employ the technique, according to the developers. The family includes 14-billion-parameter text-to-video and image-to-video models as well as a new 5-billion-parameter model that handles text-plus-image prompts. All versions produce 480p and 720p footage and aim for “cinematic-level” aesthetics while handling more complex scene composition than earlier releases. Pre-trained weights and code are available immediately on Hugging Face, GitHub and Modelscope, enabling researchers and developers to fine-tune or deploy the system without licensing fees. The launch underscores accelerating open-source competition in generative video, two years after the debut of Stable Video Diffusion and similar proprietary tools.
Sources
- Lisan al Gaib
Official release of Wan2.2 an open-source text-to-video and image-to-video model! https://t.co/3fAXZfIj56 https://t.co/qkR7QoLZoN
- Wan
🚀 Introducing Wan2.2: The World's First Open-Source MoE-Architecture Video Generation Model with Cinematic Control! 🔥 Key Innovations: ꔷ World's First Open-Source MoE Video Model: Our Mixture-of-Experts architecture scales model capacity without increasing computational
- tphuang
Videos generated by Wan 2.2. Click on them & see Text to Video model 14B Image to Video model 14B Text+Image to Video 5B Available on HF, Github & Modelscope Great stuff from Alibaba/Qwen team all week https://t.co/VbuQFNHmHd https://t.co/jurIAkLWau
Additional media
