tclf90 commited on
Commit
ac20bc1
·
verified ·
1 Parent(s): 50e9a67

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -14
README.md CHANGED
@@ -13,23 +13,20 @@ base_model_relation: quantized
13
  Base Model: [Qwen/Qwen3-VL-235B-A22B-Thinking](https://www.modelscope.cn/models/Qwen/Qwen3-VL-235B-A22B-Thinking)
14
 
15
  ### 【Dependencies / Installation】
16
- As of **2025-09-26**, create a fresh Python environment and run:
17
  ```bash
18
- pip install -U pip
19
- pip install uv
20
- pip install git+https://github.com/huggingface/transformers
21
- pip install accelerate
22
- pip install qwen-vl-utils==0.0.14
23
- # pip install 'vllm>0.10.2' # If this is not working use the below one.
24
- uv pip install -U vllm \
25
- --torch-backend=auto \
26
- --extra-index-url https://wheels.vllm.ai/nightly
27
- ```
28
- or use the docker image from qwen3vl team:
29
- ```
30
- docker run --gpus all --ipc=host --network=host --rm --name qwen3vl -it qwenllm/qwenvl:qwen3vl-cu128 bash
31
  ```
32
 
 
 
33
  ### 【vLLM Startup Command】
34
  <i>Note: When launching with TP=8, include `--enable-expert-parallel`;
35
  otherwise the expert tensors couldn’t be evenly sharded across GPU devices.</i>
 
13
  Base Model: [Qwen/Qwen3-VL-235B-A22B-Thinking](https://www.modelscope.cn/models/Qwen/Qwen3-VL-235B-A22B-Thinking)
14
 
15
  ### 【Dependencies / Installation】
16
+ As of **2025-10-08**, create a fresh Python environment and run:
17
  ```bash
18
+ uv venv
19
+ source .venv/bin/activate
20
+
21
+ # Install vLLM >=0.11.0
22
+ uv pip install -U vllm
23
+
24
+ # Install Qwen-VL utility library (recommended for offline inference)
25
+ uv pip install qwen-vl-utils==0.0.14
 
 
 
 
 
26
  ```
27
 
28
+ For more details, refer to [vLLM Official Qwen3-VL Guide](https://docs.vllm.ai/projects/recipes/en/latest/Qwen/Qwen3-VL.html)
29
+
30
  ### 【vLLM Startup Command】
31
  <i>Note: When launching with TP=8, include `--enable-expert-parallel`;
32
  otherwise the expert tensors couldn’t be evenly sharded across GPU devices.</i>