Wan (Open-source)
Wan by Alibaba’s Wan team is an open-source AI suite for generating videos from text and images. It handles motion physics, text rendering, and more—leading the VBench benchmark. Free to use under Apache 2.0.
Overview
Wan is an open-source AI video tool built by Alibaba’s Wan team. It transforms text and images into high-quality videos while handling motion dynamics, physics, and text rendering in both Chinese and English.
The best part is that it’s free AND supports Consumer-grade GPUs: The T2V-1.3B model requires only 8.19 GB VRAM, making it compatible with almost all consumer-grade GPUs. It can generate a 5-second 480P video on an RTX 4090 in about 4 minutes. Mind you, the size of the model is gigantic, dozens of gigabytes, so must have a lot of disk space available.
Wan is fully open-source under the Apache 2.0 license. No hidden fees, no subscriptions—just grab the code and start creating.
You can start using Wan right away. Check out their GitHub for installation and usage instructions. Pre-trained models are also available on Hugging Face and ModelScope for easy integration.
There's GGUF versions for Comfy UI. Wan2.2 is now supported in ComfyUI from Day 0!
Features include cinematic control, complex motion handling & precise semantics—all with MoE architecture.
Available in FP16 & FP8, with 5B models running on 8GB VRAM via auto-offloading.
Tags
Freeware Apache License 2.0 PC-based #Video & AnimationLinks
- Consistent Character
- Image-to-Video
- Inpainting (Videos)
- Motion Transfer (Body Movement Copy)
- Outpainting (Videos)
- Reference-to-Video
- Start-End Frames (Keyframes)
- Video Templates
Video Effects, Templates, Transformations
This tool offers the following AI models:
This tool is free to use when installed locally and is offered under Apache License 2.0.
People are testing the WAN 2.1 I2V model using compressed GGUF files in ComfyUI, mainly on RTX 3060 cards. Here's what they're seeing:
Performance on RTX 3060
- 416x416, 25 steps → About 9 minutes for 2 seconds of video
- 512x512, 25 steps → Around 13.5 minutes for 2 seconds
Key Resources & Process
- GGUF compressed models are up on Hugging Face
- Basic setup guide available [here]
- More details on the ComfyUI example page [here]
Hardware Tips
- Hardware used: 12GB VRAM, 48GB RAM (extra RAM helps a lot)
- Some users get by with 16-32GB RAM
Choosing Compression Levels
- Q4_0 was used, but higher levels (bigger files) give better quality
How It Compares
More stable than SkyReels. Less "melting" effect than some other tools.
[ Reddit ]
Generated on July 31, 2025:
Generated on May 23, 2025:
Generated on May 22, 2025:
Generated on May 19, 2025:
Generated on May 17, 2025:

Generated on May 17, 2025:
Generated on May 3, 2025:
Generated on February 28, 2025:
Generated on February 28, 2025:
Generated on February 28, 2025:
Generated on February 28, 2025:
Generated on February 26, 2025:
Generated on February 26, 2025:
Latest Wan (Open-source) News
August 26, 2025
WAN 2.2-S2V, an open-source cinematic speech-to-video model, is set to launch soon. The model generates video content based on audio input and is designed with a focus on visual fidelity.
July 31, 2025
WanGP 7.5 now works with Wan 2.2. Most things from Wan 2.1 still work fine like low VRAM use, extra long videos, image making, and MMaudio. It needs more RAM now since the model’s twice as big but still uses the same VRAM.
July 30, 2025
ComfyUI dropped a new patch for Wan 2.2. It cuts VAE decoding memory use by about 10%. The 5B I2V model got a big boost too. They also added new setup templates for the 14B models. You can grab the latest version on Git, Portable, or Desktop.
July 29, 2025
ComfyUI adds native Wan2.2 support.
Available in FP16 & FP8, with 5B models running on 8GB VRAM via auto-offloading.
July 28, 2025
Wan 2.2 is out now! Featuring a faster model, open-source access, and upgraded tools from Tongyi Lab at Alibaba Group. Improved prompt following, motion control, and visual detailing make it ideal for cinematic AI creation.
Useful Links
Wan 2.2 First/Last Frame Video Fast
Other
Hugging Face spaces for Wan 2.2 Fast with start and end frames capability.
Instareal WAN 2.2
Version
From the creators of Instagirl, this is the next step in their pursuit of perfect photorealism. Instareal is a specialized foundational LoRA for Wan 2.2
Daxamur's WAN 2.2 Workflows
Workflow
Workflow pre configured for the best balance for speed / quality
ComfyUI Wan2.2 Fun Inp Start-End Frame Video Generation Example
Workflow
How to use ComfyUI to complete the Wan2.2 Fun Inp start-end frame video generation example
Wan2.2 ComfyUI Official Native Workflow Example
Tutorial
Official usage guide for Alibaba Cloud Tongyi Wanxiang 2.2 video generation model in ComfyUI
Wan 2.2 with 8GB VRAM
Workflow
The OP used Wan 2.2 image-to-video GGUF model Q6 with image-to-video ligtx2v lora. They shared how they tweaked the ComfyUI default workflow.
Drone-style push‑in motion LoRA for Wan 2.1
LoRA
A new LoRA for Wan 2.1 brings realistic drone-style push‑in motion 🎥 Trained on 100 clips and refined through 40+ versions, it includes a ComfyUI workflow and is triggered by “Push‑in camera.” Perfect for adding cinematic movement to your videos.
Fast 4 steps Wan 2.1 I2V (14B) with CausVid LoRA
Version
CausVid is a distilled version of Wan 2.1 to run faster in just 4-8 steps, extracted as LoRA by Kijai and is compatible with 🧨 diffusers
WanGP by DeepBeepMeep
Version
Open and advanced large-scale video generative models for the GPU poor
ComfyUI Workflows + Guide for Wan2.1
Workflow
Using native ComfyUI nodes and using the kijai wan wrapper nodes allowing for more features.
Viral Effects LoRAs for Wan2.1 14B 480p I2V
LoRA
A wide range of open-sourced effects - such as Assassin, Jungle, Pirate Captain, Baby, Princess, Painting, Warrior, Samurai, Snow White, Bride, Mona Lisa, Zen, VIP, Puppy, Classy, and Disney Princess-all available on Hugging Face under the Apache 2.0 license trained on Wan2.1 14B I2V 480p model.
This page was last updated on August 26, 2025 at 6:13 AM