How to Run Wan2.2 Image to Video GGUF Models in ComfyUI (Low VRAM)

Table of Contents
1. Introduction
Welcome to the tutorial on How to Run Wan2.2 Image to Video GGUF Models in ComfyUI (Low VRAM). This guide is designed for users who want to leverage the powerful capabilities of the Wan2.2 model to convert images into stunning videos, even on systems with low VRAM (less than 12GB). The Wan2.2 model is a state-of-the-art image-to-video generation tool that utilizes advanced AI techniques to create high-quality videos from still images. In this tutorial, we will walk you through the entire process, from setting up the necessary environment to generating your first video.
By the end of this guide, you will have a comprehensive understanding of how to effectively use the Wan2.2 model in ComfyUI, ensuring that you can create dynamic video content without the need for high-end hardware.
2. Requirements for Wan2.2 Video Generation Model (I2V GGUF Model)
Before diving into the setup process, it’s crucial to ensure that your system meets the requirements for running the Wan2.2 Video Generation Model in ComfyUI, especially if you are working with low VRAM configurations. The GGUF model version of the Wan2.2 model is optimized for performance, allowing it to run efficiently on systems with less than 12GB of VRAM.
Requirement 1: ComfyUI Installed
To begin using ComfyUI, you’ll need to have it installed on your local system. Follow the step-by-step instructions in the article below to get ComfyUI up and running on Windows.
Requirement 2: Update ComfyUI
To ensure compatibility with the Wan2.2 GGUF models, make sure your ComfyUI installation is up to date.
-
Via ComfyUI Interface:
Click on "Manager" in the top-right corner of ComfyUI, then select "Update ComfyUI". -
For Windows Portable Users:
Navigate to the folder:
...\ComfyUI_windows_portable\update
and double-click the file named update_comfyui.bat.
Keeping ComfyUI updated ensures access to the latest features, bug fixes, and compatibility improvements.
Requirement 3: Download Wan 2.2 Model Files
Next, we need to download the required model files for Wan2.2. You can find them on the Wan2.2 Hugging Face main GGUF model page. Below is a table listing all the required files along with their corresponding folders where they should be placed.
File Name | Hugging Face Download Page | File Directory |
---|---|---|
Wan2.2-I2V-A14B-HighNoise-Q3_K_S.gguf | Download Page | ..\ComfyUI\models\diffusion_models |
Wan2.2-I2V-A14B-LowNoise-Q3_K_S.gguf | Download Page | ..\ComfyUI\models\diffusion_models |
umt5_xxl_fp8_e4m3fn_scaled.safetensors | Download Page | ..\ComfyUI\models\clip |
Wan2.1_VAE.safetensors | Download Page | ..\ComfyUI\models\vae |
lightx2v_I2V_14B_480p_cfg_step_distill_rank256_bf16.safetensors | Download Page | ..\ComfyUI\models\loras |
To wrap up, you'll also notice that we’ll be downloading a LoRA model. This helps accelerate the rendering process by allowing you to reduce the KSampler steps to a minimum—speeding up generation without sacrificing quality in the final video output.
Verify Folder Structure
Before we proceed, let’s ensure that all Wan2.2 model files are correctly placed in their respective ComfyUI folders. Having everything in the right place ensures the workflow loads smoothly without errors.
ts1📁 ComfyUI/ 2└── 📁 models/ 3 ├── 📁 clip/ 4 │ └── umt5_xxl_fp8_e4m3fn_scaled.safetensors 5 ├── 📁 diffusion_models/ 6 │ ├── Wan2.2-I2V-A14B-HighNoise-Q3_K_S.gguf 7 │ └── Wan2.2-I2V-A14B-LowNoise-Q3_K_S.gguf 8 ├── 📁 vae/ 9 │ └── Wan2.1_VAE.safetensors 10 └── 📁 loras/ 11 └── lightx2v_I2V_14B_480p_cfg_step_distill_rank256_bf16.safetensors
Now that all the required files are downloaded and properly organized, let’s move on to downloading and loading the Wan2.2 GGUF JSON Workflow so we can start generating high-quality videos from images!
3. Downloading and Loading the Wan2.2 Workflow for ComfyUI (I2V GGUF)
With the requirements in place, the next step is to download and load the Wan2.2 GGUF JSON Workflow into ComfyUI. This workflow is essential for configuring the model and ensuring it operates correctly.
Step 1: Download the Workflow File
Begin by downloading the workflow file specifically designed for the Wan2.2 GGUF model. This file contains all the necessary configurations and settings required for the model to function properly. You can find the download link below:
👉 Download Wan2.2 I2V GGUF Workflow JSON
Step 2: Load the Workflow in ComfyUI
Once you’ve downloaded the workflow file, launch ComfyUI and simply drag and drop the .json file onto the canvas to load the full setup and begin working with the I2V model.
In the next section, we’ll go over all the essential settings needed to run the Wan2.2 GGUF image to video model smoothly and efficiently.
4. Wan2.2 Video Generation Settings (I2V)
Now that the Wan2.2 GGUF workflow is loaded into ComfyUI, it’s time to configure the settings to optimize video generation. Proper configuration is crucial for achieving high-quality outputs, especially when working with low VRAM systems.
Step 1: Load Models (via Unet Loader)
Use the Unet Loader node to load both high and low noise models:
-
High Noise: Wan2.2-I2V-A14B-HighNoise-Q3_K_S.gguf
-
Low Noise: Wan2.2-I2V-A14B-LowNoise-Q3_K_S.gguf
Also load:
-
CLIP Text Encoder: umt5_xxl_fp8_e4m3fn_scaled.safetensors
-
VAE: Wan2.1_VAE.safetensors
Step 2: Upload Base Image
Import the base image that will act as the first frame of the video. This defines character design, pose, lighting, and scene layout.
Step 3: Load LoRA
Use a LoRA Loader node to load:
- lightx2v_I2V_14B_480p_cfg_step_distill_rank256_bf16.safetensors
This LoRA speeds up rendering without sacrificing quality.
Step 4: Enter Prompt and Negative Prompt
Write a detailed prompt to describe the animation behavior and camera motion.
Step 5: Set Video Resolution and Length
Match your base image aspect ratio and define animation duration.
-
Width: 1280
-
Height: 720
-
Length: 121 frames
-
Batch Size: 1
Step 6: Configure High Noise Ksampler Settings
Settings for motion-rich, expressive animation:
-
Add Noise: Enabled
-
Control After Generate: Randomize
-
Total Steps: 6
-
CFG: 1
-
Sampler: Euler
-
Scheduler: Simple
-
Start at Step: 0
-
Start end Step: 3
-
Return with Leftover Noise: Enabled
Step 7: Configure Low Noise Ksampler Settings
Stabilizes animation and refines details:
-
Add Noise: Disabled
-
Control After Generate: Randomize
-
Total Steps: 6
-
CFG: 1
-
Sampler: Euler
-
Scheduler: Simple
-
Start at Step: 3
-
Start end Step: 10000
-
Return with Leftover Noise: Disabled
Step 8: Set FPS and Video Output Settings
-
FPS: 30
Creates a smooth ~4-second clip (121 ÷ 30 ≈ 4s). -
Configure the video output:
-
File Prefix: videoComfyUI
-
Format: auto or mp4
-
Codec: h264
-
With everything in place, click RUN to generate your animated video.
5. Wan2.2 Video Generation Example
After configuring the settings and clicking "RUN", the Wan2.2 model will begin generating your video. Depending on your hardware capabilities, the generation time may vary, especially if you are working with longer or higher-resolution videos.
Example 1:
Example 2:
Tips for Success
-
Start with shorter video lengths to gauge performance on your system.
-
Experiment with various prompts to see how they influence the final output.
-
Keep an eye on the resource usage to ensure smooth operation on low VRAM systems.
With these examples and tips, you are now ready to explore the capabilities of the Wan2.2 GGUF image-to-video model in ComfyUI.
6. Conclusion
This tutorial has guided you through running the Wan2.2 image-to-video GGUF models in ComfyUI, optimized for low VRAM usage. From model setup to workflow configuration and video generation, you now have a complete understanding of how to bring still images to life.
By integrating the lightx2v LoRA, we've significantly reduced rendering time while maintaining visual quality—making the process more efficient and accessible, even on modest hardware.
As AI tools continue to evolve, we encourage you to experiment, refine, and stay up to date with new improvements. Thanks for following along—now go create something amazing!