On February 25, 2025, Alibaba's Tongyi Lab launched Wan 2.1, an open-source video generation suite that can be accessed on Hugging Face and allows text-to-video production using consumer grade hardware. As of March 13, 2025, this paper explains how to install the Wan 2.1 T2V-1.3B model locally on a low cost GPU, like an NVIDIA RTX 4090, to produce a 5-second 480p video. It caters to developers and enthusiasts looking for easily available AI video tools in India or around the world.
Step 1: Install Dependencies
Update system and install essentials:
sudo apt update && sudo apt install -y python3-pip git
pip3 install torch torchvision --index-url https://download.pytorch.org/whl/cu118
Step 2: Download Wan 2.1 from Hugging Face
Clone the GitHub repo and download the 1.3B model:
git clone https://github.com/Wan-Video/Wan2.1.git
cd Wan2.1
huggingface-cli download Wan-AI/Wan2.1-T2V-1.3B --local-dir ./models
Step 3: Install Requirements
Install dependencies from the repo:
pip3 install -r requirements.txt
Step 4: Generate a Video
Run the inference script with a text prompt:
python3 inference.py --model_path ./models/Wan2.1-T2V-1.3B --prompt "A cat jumping over a fence" --output video.mp4 --resolution 480p --duration 5
Output: video.mp4 (5 seconds, 480p).
Minimum System Requirements
Deploying Wan 2.1 on a local server with a budget GPU like the RTX 4090 is straightforward, leveraging its Hugging Face availability. This setup empowers Indian developers and hobbyists to generate 480p videos in minutes, showcasing accessible AI innovation. Wan 2.1’s lightweight 1.3B model makes video creation a reality without breaking the bank.
Ready to transform your business with our technology solutions? Contact Us today to Leverage Our AI/ML Expertise.