noctrex commited on
Commit
d8e63f6
·
verified ·
1 Parent(s): 8371d85

Add files using upload-large-folder tool

Browse files
.gitattributes CHANGED
@@ -34,3 +34,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  Qwen3-Next-80B-A3B-Thinking-MXFP4_MOE.gguf filter=lfs diff=lfs merge=lfs -text
 
 
 
 
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  Qwen3-Next-80B-A3B-Thinking-MXFP4_MOE.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Qwen3-Next-80B-A3B-Thinking-MXFP4_MOE-00001-of-00003.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Qwen3-Next-80B-A3B-Thinking-MXFP4_MOE-00002-of-00003.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Qwen3-Next-80B-A3B-Thinking-MXFP4_MOE-00003-of-00003.gguf filter=lfs diff=lfs merge=lfs -text
Qwen3-Next-80B-A3B-Thinking-MXFP4_MOE-00001-of-00003.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98c6429c9cf562c4f3086e111de20b6cee82767e5aac4474a882c81b5c41d5a9
3
+ size 14738984864
Qwen3-Next-80B-A3B-Thinking-MXFP4_MOE-00002-of-00003.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eca385a3882809860f50f5e27a55d2a06d23ca08891c333c0a36069d8bb56248
3
+ size 14930823232
Qwen3-Next-80B-A3B-Thinking-MXFP4_MOE-00003-of-00003.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e340867bc43202e0bc0c38718e86b6e95a0cf097824849bf6fd4b7f3fde2012
3
+ size 14071817696
README.md CHANGED
@@ -1,35 +1,29 @@
1
- ---
2
- base_model:
3
- - Qwen/Qwen3-Next-80B-A3B-Thinking
4
- pipeline_tag: text-generation
5
- ---
6
- This is a MXFP4 quant of Qwen3-Next-80B-A3B-Thinking
7
-
8
- Welcome to the bleeding edge.
9
-
10
- I must to point out that this a *experimental* release.
11
-
12
- Say it after me, *EXPERIMENTAL*.
13
-
14
- This has been made possible cause of the excellent work done by [pwilkin](https://github.com/pwilkin/llama.cpp) and others.
15
-
16
- He has a development branch of llama.cpp for Qwen3-Next.
17
-
18
- It has not yet been released officially, and things are moving quite fast.
19
-
20
- For the time being, as of 2025-10-24 I got the source code from his fork and compiled in order to be able to generate the GGUF's, from here:
21
- https://github.com/pwilkin/llama.cpp/tree/qwen3_next
22
-
23
- This GGUF will run only with it.
24
-
25
- If you cannot compile it yourself, I have made a Windows version with Vulkan support you can find here:
26
-
27
- [llama-qwen3-next-5edfe78-bin-win-vulkan-x64.zip](https://gofile.io/d/qhHL6n)
28
-
29
- I should state that this may trigger false positives from your AV, this has NO virus, I compiled it on a my Windows 11 PC that i check regularly for viruses.
30
-
31
- If you don't trust strangers giving out binaries, you can try compiling it for yourself, in order to be sure.
32
-
33
- https://www.virustotal.com/gui/file/35a134a8977488ff6b82ce3f2b5df20da742ec212859a5e0c30813c55519f4f0
34
-
35
- When the support for Qwen3-Next officially releases on mainline llama.cpp, I will see if these files will need a new updated quantization, and update if needed.
 
1
+ ---
2
+ pipeline_tag: text-generation
3
+ base_model:
4
+ - Qwen/Qwen3-Next-80B-A3B-Thinking
5
+ ---
6
+ This is a MXFP4 quant of Qwen3-Next-80B-A3B-Thinking
7
+
8
+ Welcome to the bleeding edge.
9
+ I must to point out that this a *experimental* release.
10
+ Say it after me, *EXPERIMENTAL*.
11
+
12
+ This has been made possible cause of the excellent work done by [pwilkin](https://github.com/pwilkin/llama.cpp) and others.
13
+
14
+ He has a development branch of llama.cpp for Qwen3-Next.
15
+ It has not yet been released officially, and things are moving quite fast.
16
+
17
+ For the time being, as of 2025-10-24 I got the source code from his fork and compiled in order to be able to generate the GGUF's, from here:
18
+ https://github.com/pwilkin/llama.cpp/tree/qwen3_next
19
+
20
+ This GGUF will run only with it.
21
+
22
+ If you cannot compile it yourself, I have made a Windows version with Vulkan support you can find here:
23
+ [llama-qwen3-next-5edfe78-bin-win-vulkan-x64.zip](https://gofile.io/d/qhHL6n)
24
+ I should state that this may trigger false positives from your AV, this has NO virus, I compiled it on a my Windows 11 PC that i check regularly for viruses.
25
+ If you don't trust strangers giving out binaries, you can try compiling it for yourself, in order to be sure.
26
+
27
+ https://www.virustotal.com/gui/file/35a134a8977488ff6b82ce3f2b5df20da742ec212859a5e0c30813c55519f4f0
28
+
29
+ When the support for Qwen3-Next officially releases on mainline llama.cpp, I will see if these files will need a new updated quantization, and update if needed.