Spaces:
Running
Running
# Troubleshooting Guide | |
This guide covers common issues and their solutions when using WanGP. | |
## Installation Issues | |
### PyTorch Installation Problems | |
#### CUDA Version Mismatch | |
**Problem**: PyTorch can't detect GPU or CUDA errors | |
**Solution**: | |
```bash | |
# Check your CUDA version | |
nvidia-smi | |
# Install matching PyTorch version | |
# For CUDA 12.4 (RTX 10XX-40XX) | |
pip install torch==2.6.0 torchvision torchaudio --index-url https://download.pytorch.org/whl/test/cu124 | |
# For CUDA 12.8 (RTX 50XX) | |
pip install torch==2.7.0 torchvision torchaudio --index-url https://download.pytorch.org/whl/test/cu128 | |
``` | |
#### Python Version Issues | |
**Problem**: Package compatibility errors | |
**Solution**: Ensure you're using Python 3.10.9 | |
```bash | |
python --version # Should show 3.10.9 | |
conda create -n wan2gp python=3.10.9 | |
``` | |
### Dependency Installation Failures | |
#### Triton Installation (Windows) | |
**Problem**: `pip install triton-windows` fails | |
**Solution**: | |
1. Update pip: `pip install --upgrade pip` | |
2. Try pre-compiled wheel | |
3. Fallback to SDPA attention: `python wgp.py --attention sdpa` | |
#### SageAttention Compilation Issues | |
**Problem**: SageAttention installation fails | |
**Solution**: | |
1. Install Visual Studio Build Tools (Windows) | |
2. Use pre-compiled wheels when available | |
3. Fallback to basic attention modes | |
## Memory Issues | |
### CUDA Out of Memory | |
#### During Model Loading | |
**Problem**: "CUDA out of memory" when loading model | |
**Solutions**: | |
```bash | |
# Use smaller model | |
python wgp.py --t2v-1-3B | |
# Enable quantization (usually default) | |
python wgp.py --quantize-transformer True | |
# Use memory-efficient profile | |
python wgp.py --profile 4 | |
# Reduce preloaded model size | |
python wgp.py --preload 0 | |
``` | |
#### During Video Generation | |
**Problem**: Memory error during generation | |
**Solutions**: | |
1. Reduce frame count (shorter videos) | |
2. Lower resolution in advanced settings | |
3. Use lower batch size | |
4. Clear GPU cache between generations | |
### System RAM Issues | |
#### High RAM Usage | |
**Problem**: System runs out of RAM | |
**Solutions**: | |
```bash | |
# Limit reserved memory | |
python wgp.py --perc-reserved-mem-max 0.3 | |
# Use minimal RAM profile | |
python wgp.py --profile 5 | |
# Enable swap file (OS level) | |
``` | |
## Performance Issues | |
### Slow Generation Speed | |
#### General Optimization | |
```bash | |
# Enable compilation (requires Triton) | |
python wgp.py --compile | |
# Use faster attention | |
python wgp.py --attention sage2 | |
# Enable TeaCache | |
python wgp.py --teacache 2.0 | |
# Use high-performance profile | |
python wgp.py --profile 3 | |
``` | |
#### GPU-Specific Optimizations | |
**RTX 10XX/20XX Series**: | |
```bash | |
python wgp.py --attention sdpa --profile 4 --teacache 1.5 | |
``` | |
**RTX 30XX/40XX Series**: | |
```bash | |
python wgp.py --compile --attention sage --profile 3 --teacache 2.0 | |
``` | |
**RTX 50XX Series**: | |
```bash | |
python wgp.py --attention sage --profile 4 --fp16 | |
``` | |
### Attention Mechanism Issues | |
#### Sage Attention Not Working | |
**Problem**: Sage attention fails to compile or work | |
**Diagnostic Steps**: | |
1. Check Triton installation: | |
```python | |
import triton | |
print(triton.__version__) | |
``` | |
2. Clear Triton cache: | |
```bash | |
# Windows | |
rmdir /s %USERPROFILE%\.triton | |
# Linux | |
rm -rf ~/.triton | |
``` | |
3. Fallback solution: | |
```bash | |
python wgp.py --attention sdpa | |
``` | |
#### Flash Attention Issues | |
**Problem**: Flash attention compilation fails | |
**Solution**: | |
- Windows: Often requires manual CUDA kernel compilation | |
- Linux: Usually works with `pip install flash-attn` | |
- Fallback: Use Sage or SDPA attention | |
## Model-Specific Issues | |
### Lora Problems | |
#### Loras Not Loading | |
**Problem**: Loras don't appear in the interface | |
**Solutions**: | |
1. Check file format (should be .safetensors, .pt, or .pth) | |
2. Verify correct directory: | |
``` | |
loras/ # For t2v models | |
loras_i2v/ # For i2v models | |
loras_hunyuan/ # For Hunyuan models | |
``` | |
3. Click "Refresh" button in interface | |
4. Use `--check-loras` to filter incompatible files | |
#### Lora Compatibility Issues | |
**Problem**: Lora causes errors or poor results | |
**Solutions**: | |
1. Check model size compatibility (1.3B vs 14B) | |
2. Verify lora was trained for your model type | |
3. Try different multiplier values | |
4. Use `--check-loras` flag to auto-filter | |
### VACE-Specific Issues | |
#### Poor VACE Results | |
**Problem**: VACE generates poor quality or unexpected results | |
**Solutions**: | |
1. Enable Skip Layer Guidance | |
2. Use detailed prompts describing all elements | |
3. Ensure proper mask creation with Matanyone | |
4. Check reference image quality | |
5. Use at least 15 steps, preferably 30+ | |
#### Matanyone Tool Issues | |
**Problem**: Mask creation difficulties | |
**Solutions**: | |
1. Use negative point prompts to refine selection | |
2. Create multiple sub-masks and combine them | |
3. Try different background removal options | |
4. Ensure sufficient contrast in source video | |
## Network and Server Issues | |
### Gradio Interface Problems | |
#### Port Already in Use | |
**Problem**: "Port 7860 is already in use" | |
**Solution**: | |
```bash | |
# Use different port | |
python wgp.py --server-port 7861 | |
# Or kill existing process | |
# Windows | |
netstat -ano | findstr :7860 | |
taskkill /PID <PID> /F | |
# Linux | |
lsof -i :7860 | |
kill <PID> | |
``` | |
#### Interface Not Loading | |
**Problem**: Browser shows "connection refused" | |
**Solutions**: | |
1. Check if server started successfully | |
2. Try `http://127.0.0.1:7860` instead of `localhost:7860` | |
3. Disable firewall temporarily | |
4. Use `--listen` flag for network access | |
### Remote Access Issues | |
#### Sharing Not Working | |
**Problem**: `--share` flag doesn't create public URL | |
**Solutions**: | |
1. Check internet connection | |
2. Try different network | |
3. Use `--listen` with port forwarding | |
4. Check firewall settings | |
## Quality Issues | |
### Poor Video Quality | |
#### General Quality Improvements | |
1. Increase number of steps (25-30+) | |
2. Use larger models (14B instead of 1.3B) | |
3. Enable Skip Layer Guidance | |
4. Improve prompt descriptions | |
5. Use higher resolution settings | |
#### Specific Quality Issues | |
**Blurry Videos**: | |
- Increase steps | |
- Check source image quality (i2v) | |
- Reduce TeaCache multiplier | |
- Use higher guidance scale | |
**Inconsistent Motion**: | |
- Use longer overlap in sliding windows | |
- Reduce window size | |
- Improve prompt consistency | |
- Check control video quality (VACE) | |
**Color Issues**: | |
- Check model compatibility | |
- Adjust guidance scale | |
- Verify input image color space | |
- Try different VAE settings | |
## Advanced Debugging | |
### Enable Verbose Output | |
```bash | |
# Maximum verbosity | |
python wgp.py --verbose 2 | |
# Check lora compatibility | |
python wgp.py --check-loras --verbose 2 | |
``` | |
### Memory Debugging | |
```bash | |
# Monitor GPU memory | |
nvidia-smi -l 1 | |
# Reduce memory usage | |
python wgp.py --profile 4 --perc-reserved-mem-max 0.2 | |
``` | |
### Performance Profiling | |
```bash | |
# Test different configurations | |
python wgp.py --attention sdpa --profile 4 # Baseline | |
python wgp.py --attention sage --profile 3 # Performance | |
python wgp.py --compile --teacache 2.0 # Maximum speed | |
``` | |
## Getting Help | |
### Before Asking for Help | |
1. Check this troubleshooting guide | |
2. Read the relevant documentation: | |
- [Installation Guide](INSTALLATION.md) | |
- [Getting Started](GETTING_STARTED.md) | |
- [Command Line Reference](CLI.md) | |
3. Try basic fallback configuration: | |
```bash | |
python wgp.py --attention sdpa --profile 4 | |
``` | |
### Community Support | |
- **Discord Server**: https://discord.gg/g7efUW9jGV | |
- Provide relevant information: | |
- GPU model and VRAM amount | |
- Python and PyTorch versions | |
- Complete error messages | |
- Command used to launch WanGP | |
- Operating system | |
### Reporting Bugs | |
When reporting issues: | |
1. Include system specifications | |
2. Provide complete error logs | |
3. List the exact steps to reproduce | |
4. Mention any modifications to default settings | |
5. Include command line arguments used | |
## Emergency Fallback | |
If nothing works, try this minimal configuration: | |
```bash | |
# Absolute minimum setup | |
python wgp.py --t2v-1-3B --attention sdpa --profile 4 --teacache 0 --fp16 | |
# If that fails, check basic PyTorch installation | |
python -c "import torch; print(torch.cuda.is_available())" | |
``` |