Add files using upload-large-folder tool
Browse files
.gitattributes
CHANGED
|
@@ -34,3 +34,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
Qwen3-Next-80B-A3B-Thinking-MXFP4_MOE.gguf filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
Qwen3-Next-80B-A3B-Thinking-MXFP4_MOE.gguf filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
Qwen3-Next-80B-A3B-Thinking-MXFP4_MOE-00001-of-00003.gguf filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
Qwen3-Next-80B-A3B-Thinking-MXFP4_MOE-00002-of-00003.gguf filter=lfs diff=lfs merge=lfs -text
|
| 39 |
+
Qwen3-Next-80B-A3B-Thinking-MXFP4_MOE-00003-of-00003.gguf filter=lfs diff=lfs merge=lfs -text
|
Qwen3-Next-80B-A3B-Thinking-MXFP4_MOE-00001-of-00003.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:98c6429c9cf562c4f3086e111de20b6cee82767e5aac4474a882c81b5c41d5a9
|
| 3 |
+
size 14738984864
|
Qwen3-Next-80B-A3B-Thinking-MXFP4_MOE-00002-of-00003.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eca385a3882809860f50f5e27a55d2a06d23ca08891c333c0a36069d8bb56248
|
| 3 |
+
size 14930823232
|
Qwen3-Next-80B-A3B-Thinking-MXFP4_MOE-00003-of-00003.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4e340867bc43202e0bc0c38718e86b6e95a0cf097824849bf6fd4b7f3fde2012
|
| 3 |
+
size 14071817696
|
README.md
CHANGED
|
@@ -1,35 +1,29 @@
|
|
| 1 |
-
---
|
| 2 |
-
|
| 3 |
-
|
| 4 |
-
|
| 5 |
-
---
|
| 6 |
-
This is a MXFP4 quant of Qwen3-Next-80B-A3B-Thinking
|
| 7 |
-
|
| 8 |
-
Welcome to the bleeding edge.
|
| 9 |
-
|
| 10 |
-
|
| 11 |
-
|
| 12 |
-
|
| 13 |
-
|
| 14 |
-
|
| 15 |
-
|
| 16 |
-
|
| 17 |
-
|
| 18 |
-
|
| 19 |
-
|
| 20 |
-
|
| 21 |
-
|
| 22 |
-
|
| 23 |
-
|
| 24 |
-
|
| 25 |
-
If you
|
| 26 |
-
|
| 27 |
-
|
| 28 |
-
|
| 29 |
-
|
| 30 |
-
|
| 31 |
-
If you don't trust strangers giving out binaries, you can try compiling it for yourself, in order to be sure.
|
| 32 |
-
|
| 33 |
-
https://www.virustotal.com/gui/file/35a134a8977488ff6b82ce3f2b5df20da742ec212859a5e0c30813c55519f4f0
|
| 34 |
-
|
| 35 |
-
When the support for Qwen3-Next officially releases on mainline llama.cpp, I will see if these files will need a new updated quantization, and update if needed.
|
|
|
|
| 1 |
+
---
|
| 2 |
+
pipeline_tag: text-generation
|
| 3 |
+
base_model:
|
| 4 |
+
- Qwen/Qwen3-Next-80B-A3B-Thinking
|
| 5 |
+
---
|
| 6 |
+
This is a MXFP4 quant of Qwen3-Next-80B-A3B-Thinking
|
| 7 |
+
|
| 8 |
+
Welcome to the bleeding edge.
|
| 9 |
+
I must to point out that this a *experimental* release.
|
| 10 |
+
Say it after me, *EXPERIMENTAL*.
|
| 11 |
+
|
| 12 |
+
This has been made possible cause of the excellent work done by [pwilkin](https://github.com/pwilkin/llama.cpp) and others.
|
| 13 |
+
|
| 14 |
+
He has a development branch of llama.cpp for Qwen3-Next.
|
| 15 |
+
It has not yet been released officially, and things are moving quite fast.
|
| 16 |
+
|
| 17 |
+
For the time being, as of 2025-10-24 I got the source code from his fork and compiled in order to be able to generate the GGUF's, from here:
|
| 18 |
+
https://github.com/pwilkin/llama.cpp/tree/qwen3_next
|
| 19 |
+
|
| 20 |
+
This GGUF will run only with it.
|
| 21 |
+
|
| 22 |
+
If you cannot compile it yourself, I have made a Windows version with Vulkan support you can find here:
|
| 23 |
+
[llama-qwen3-next-5edfe78-bin-win-vulkan-x64.zip](https://gofile.io/d/qhHL6n)
|
| 24 |
+
I should state that this may trigger false positives from your AV, this has NO virus, I compiled it on a my Windows 11 PC that i check regularly for viruses.
|
| 25 |
+
If you don't trust strangers giving out binaries, you can try compiling it for yourself, in order to be sure.
|
| 26 |
+
|
| 27 |
+
https://www.virustotal.com/gui/file/35a134a8977488ff6b82ce3f2b5df20da742ec212859a5e0c30813c55519f4f0
|
| 28 |
+
|
| 29 |
+
When the support for Qwen3-Next officially releases on mainline llama.cpp, I will see if these files will need a new updated quantization, and update if needed.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|