Alibaba Releases Wan2.1 Free AI Video Generation Tool

Alibaba has made its Wan2.1 series video generation AI models available free of charge. The open-source release not only boosts global accessibility and accelerates innovation but also outperforms competitors on industry benchmarks—solidifying Alibaba’s leadership in the rapidly evolving AI market.

Alibaba Launches Wan2.1 Free AI Video Generation Tool - Credit - The AI Track, Alibaba, Canva
Alibaba Launches Wan2.1 Free AI Video Generation Tool - Credit - The AI Track, Alibaba, Canva

Article – Key Points

  • Open Sourcing Wan2.1 Models:

    On February 26, 2025, Alibaba announced that four variants from its Wan2.1 series are now free for worldwide use. These include:

    • T2V-1.3B: A smaller model designed for generating a five-second 480P video using only 8.19 GB of VRAM on an RTX 4090 GPU.
    • T2V-14B: Capable of generating videos up to 720P and built with 14 billion parameters to process extensive inputs for highly accurate outputs.
    • I2V-14B-720P and I2V-14B-480P: Models that generate images and videos from text and image inputs at different resolutions.

    Distributed via Alibaba Cloud’s ModelScope and Hugging Face, these models are accessible to academics, researchers, and commercial institutions globally.

  • Market Impact:

    Following the announcement, Alibaba’s Hong Kong-listed shares increased nearly 5% on the day, contributing to a 66% surge in its stock value in 2025. This robust market performance reflects Alibaba’s improved financial metrics and growing influence as a key AI player in China. Furthermore, Alibaba announced plans to invest at least 380 billion yuan (approximately $52 billion) over the next three years to bolster its cloud computing and AI infrastructure, reinforcing its long-term commitment to technological innovation.

  • Competitive Landscape:

    Alibaba’s open-source strategy intensifies competition in the AI market. The Wan2.1 models have achieved top rankings on the VBench leaderboard—evaluating criteria such as motion smoothness, text rendering, physics accuracy, and multi-object interaction—surpassing rivals like OpenAI’s Sora, Hailuo’s MiniMax, and Luma’s Dream Machine. This move follows similar initiatives by Chinese startup DeepSeek, whose cost-effective open-source models have already generated significant excitement among investors and industry observers. By releasing both the code and model weights, Alibaba empowers developers to create customized spinoffs, in contrast to the proprietary systems favored by competitors such as OpenAI and Meta.

  • Technological Advancements:

    The Wan2.1 series is built on advanced technologies that include a novel spatio-temporal variational autoencoder (VAE) capable of reconstructing video at 2.5 times the speed of competing systems on an A800 GPU. It leverages scalable pre-training strategies, employs a Flow Matching framework within the Diffusion Transformer (DiT) paradigm, and integrates a T5 encoder for robust multi-language text processing. The training pipeline involved curating and deduplicating 1.5 billion videos and 10 billion images, ensuring high-quality data for robust video generation. This suite of innovations places Wan2.1 at the forefront of the industry.

  • Historical Strategy and Evolution:

    Alibaba’s journey into open-source AI began in August 2023 when it launched its first open-source model. In January 2025, the company introduced the latest version of its video- and image-generating AI model—shortening its name from Wanx to Wan. The release of Wan2.1 marks a significant escalation in this strategy, democratizing access to advanced AI and fostering a community-driven innovation ecosystem that challenges the entrenched dominance of proprietary systems.

  • Global Accessibility and Future Trends:

    By lowering the barriers to AI research and commercial use, Alibaba’s free release of its Wan2.1 models is poised to accelerate global technological innovation. The models support a wide range of functionalities—including text-to-video, image-to-video, text-to-image, video editing, and video-to-audio—which will likely set new industry standards. Notably, the T2V-14B variant supports both Chinese and English text, broadening its appeal and usability on a global scale. Additionally, on the same week, Alibaba previewed its new reasoning model QwQ-Max, which it plans to release as open source upon full rollout, further underscoring its commitment to fostering innovation.


Why This Matters:

Alibaba’s decision to open source its Wan2.1 AI video generation models marks a pivotal shift in the global AI landscape. By making cutting-edge models accessible to everyone, Alibaba is setting new standards for innovation and competition. This move not only empowers researchers and developers to build on Alibaba’s technology but also pressures competitors to reexamine their proprietary approaches. In an era where AI is central to technological advancement, such open-source initiatives are likely to drive faster innovation, reduce development costs, and create more equitable access to state-of-the-art tools—benefiting industries and consumers worldwide.

AI is revolutionizing filmmaking and content creation! This comprehensive guide compares the top 20 text-to-video tools, highlighting their strengths, and limitations

Read a comprehensive monthly roundup of the latest AI news!

The AI Track News: In-Depth And Concise

Scroll to Top