Spaces:
Running
Running
--vace-1-3B--vace-1-3B# Command Line Reference | |
This document covers all available command line options for WanGP. | |
## Basic Usage | |
```bash | |
# Default launch | |
python wgp.py | |
# Specific model modes | |
python wgp.py --i2v # Image-to-video | |
python wgp.py --t2v # Text-to-video (default) | |
python wgp.py --t2v-14B # 14B text-to-video model | |
python wgp.py --t2v-1-3B # 1.3B text-to-video model | |
python wgp.py --i2v-14B # 14B image-to-video model | |
python wgp.py --i2v-1-3B # Fun InP 1.3B image-to-video model | |
python wgp.py --vace-1-3B # VACE ControlNet 1.3B model | |
``` | |
## Model and Performance Options | |
### Model Configuration | |
```bash | |
--quantize-transformer BOOL # Enable/disable transformer quantization (default: True) | |
--compile # Enable PyTorch compilation (requires Triton) | |
--attention MODE # Force attention mode: sdpa, flash, sage, sage2 | |
--profile NUMBER # Performance profile 1-5 (default: 4) | |
--preload NUMBER # Preload N MB of diffusion model in VRAM | |
--fp16 # Force fp16 instead of bf16 models | |
--gpu DEVICE # Run on specific GPU device (e.g., "cuda:1") | |
``` | |
### Performance Profiles | |
- **Profile 1**: Load entire current model in VRAM and keep all unused models in reserved RAM for fast VRAM tranfers | |
- **Profile 2**: Load model parts as needed, keep all unused models in reserved RAM for fast VRAM tranfers | |
- **Profile 3**: Load entire current model in VRAM (requires 24GB for 14B model) | |
- **Profile 4**: Default and recommended, load model parts as needed, most flexible option | |
- **Profile 5**: Minimum RAM usage | |
### Memory Management | |
```bash | |
--perc-reserved-mem-max FLOAT # Max percentage of RAM for reserved memory (< 0.5) | |
``` | |
## Lora Configuration | |
```bash | |
--lora-dir PATH # Path to Wan t2v loras directory | |
--lora-dir-i2v PATH # Path to Wan i2v loras directory | |
--lora-dir-hunyuan PATH # Path to Hunyuan t2v loras directory | |
--lora-dir-hunyuan-i2v PATH # Path to Hunyuan i2v loras directory | |
--lora-dir-ltxv PATH # Path to LTX Video loras directory | |
--lora-preset PRESET # Load lora preset file (.lset) on startup | |
--check-loras # Filter incompatible loras (slower startup) | |
``` | |
## Generation Settings | |
### Basic Generation | |
```bash | |
--seed NUMBER # Set default seed value | |
--frames NUMBER # Set default number of frames to generate | |
--steps NUMBER # Set default number of denoising steps | |
--advanced # Launch with advanced mode enabled | |
``` | |
### Advanced Generation | |
```bash | |
--teacache MULTIPLIER # TeaCache speed multiplier: 0, 1.5, 1.75, 2.0, 2.25, 2.5 | |
``` | |
## Interface and Server Options | |
### Server Configuration | |
```bash | |
--server-port PORT # Gradio server port (default: 7860) | |
--server-name NAME # Gradio server name (default: localhost) | |
--listen # Make server accessible on network | |
--share # Create shareable HuggingFace URL for remote access | |
--open-browser # Open browser automatically when launching | |
``` | |
### Interface Options | |
```bash | |
--lock-config # Prevent modifying video engine configuration from interface | |
--theme THEME_NAME # UI theme: "default" or "gradio" | |
``` | |
## File and Directory Options | |
```bash | |
--settings PATH # Path to folder containing default settings for all models | |
--verbose LEVEL # Information level 0-2 (default: 1) | |
``` | |
## Examples | |
### Basic Usage Examples | |
```bash | |
# Launch with specific model and loras | |
python wgp.py --t2v-14B --lora-preset mystyle.lset | |
# High-performance setup with compilation | |
python wgp.py --compile --attention sage2 --profile 3 | |
# Low VRAM setup | |
python wgp.py --t2v-1-3B --profile 4 --attention sdpa | |
# Multiple images with custom lora directory | |
python wgp.py --i2v --multiple-images --lora-dir /path/to/shared/loras | |
``` | |
### Server Configuration Examples | |
```bash | |
# Network accessible server | |
python wgp.py --listen --server-port 8080 | |
# Shareable server with custom theme | |
python wgp.py --share --theme gradio --open-browser | |
# Locked configuration for public use | |
python wgp.py --lock-config --share | |
``` | |
### Advanced Performance Examples | |
```bash | |
# Maximum performance (requires high-end GPU) | |
python wgp.py --compile --attention sage2 --profile 3 --preload 2000 | |
# Optimized for RTX 2080Ti | |
python wgp.py --profile 4 --attention sdpa --teacache 2.0 | |
# Memory-efficient setup | |
python wgp.py --fp16 --profile 4 --perc-reserved-mem-max 0.3 | |
``` | |
### TeaCache Configuration | |
```bash | |
# Different speed multipliers | |
python wgp.py --teacache 1.5 # 1.5x speed, minimal quality loss | |
python wgp.py --teacache 2.0 # 2x speed, some quality loss | |
python wgp.py --teacache 2.5 # 2.5x speed, noticeable quality loss | |
python wgp.py --teacache 0 # Disable TeaCache | |
``` | |
## Attention Modes | |
### SDPA (Default) | |
```bash | |
python wgp.py --attention sdpa | |
``` | |
- Available by default with PyTorch | |
- Good compatibility with all GPUs | |
- Moderate performance | |
### Sage Attention | |
```bash | |
python wgp.py --attention sage | |
``` | |
- Requires Triton installation | |
- 30% faster than SDPA | |
- Small quality cost | |
### Sage2 Attention | |
```bash | |
python wgp.py --attention sage2 | |
``` | |
- Requires Triton and SageAttention 2.x | |
- 40% faster than SDPA | |
- Best performance option | |
### Flash Attention | |
```bash | |
python wgp.py --attention flash | |
``` | |
- May require CUDA kernel compilation | |
- Good performance | |
- Can be complex to install on Windows | |
## Troubleshooting Command Lines | |
### Fallback to Basic Setup | |
```bash | |
# If advanced features don't work | |
python wgp.py --attention sdpa --profile 4 --fp16 | |
``` | |
### Debug Mode | |
```bash | |
# Maximum verbosity for troubleshooting | |
python wgp.py --verbose 2 --check-loras | |
``` | |
### Memory Issue Debugging | |
```bash | |
# Minimal memory usage | |
python wgp.py --profile 4 --attention sdpa --perc-reserved-mem-max 0.2 | |
``` | |
## Configuration Files | |
### Settings Files | |
Load custom settings: | |
```bash | |
python wgp.py --settings /path/to/settings/folder | |
``` | |
### Lora Presets | |
Create and share lora configurations: | |
```bash | |
# Load specific preset | |
python wgp.py --lora-preset anime_style.lset | |
# With custom lora directory | |
python wgp.py --lora-preset mystyle.lset --lora-dir /shared/loras | |
``` | |
## Environment Variables | |
While not command line options, these environment variables can affect behavior: | |
- `CUDA_VISIBLE_DEVICES` - Limit visible GPUs | |
- `PYTORCH_CUDA_ALLOC_CONF` - CUDA memory allocation settings | |
- `TRITON_CACHE_DIR` - Triton cache directory (for Sage attention) |