--- license: apache-2.0 base_model: - Qwen/Qwen3-235B-A22B-Instruct-2507 --- # Qwen3 235B Instruct 2507 gguf Make sure you have enough ram/gpu to run. On the right of model card, you may see the size of each quantized models. ## Use the model in LM Studio ### download and install LM Studio https://lmstudio.ai/ ## Discover models ### In the LM Studio, click "Discover" icon. "Mission Control" popup window will be displayed. ### In the "Mission Control" search bar, type "John1604/Qwen3-235B-A22B-Instruct-2507-gguf" and check "GGUF", the model should be found. ### Download the model. you may choose the quantized type in the download window. ### Load the model. Load the model to the LM Studio. ### Ask questions. ## quantized models | Type | Bits | Quality | Description | | ---------- | ----- | ---------------- | ------------------------------------ | | **Q2_K** | 2-bit | 🟥 Low | Minimal footprint; only for tests | | **Q3_K_S** | 3-bit | 🟧 Low | “Small” variant (less accurate) | | **Q3_K_M** | 3-bit | 🟧 Low–Med | “Medium” variant | | **Q4_K_S** | 4-bit | 🟨 Med | Small, faster, slightly less quality | | **Q4_K_M** | 4-bit | 🟩 Med–High | “Medium” — best 4-bit balance | | **Q6_K** | 6-bit | 🟩🟩🟩 Very High | Almost FP16 quality, larger size | | **Q8_0** | 8-bit | 🟩🟩🟩🟩 | Near-lossless baseline |