Text Generation
MLX
Safetensors
qwen3_moe
programming
code generation
code
codeqwen
Mixture of Experts
coding
coder
qwen2
chat
qwen
qwen-coder
Qwen3-Coder-30B-A3B-Instruct
Qwen3-30B-A3B
mixture of experts
128 experts
8 active experts
1 million context
qwen3
finetune
brainstorm 20x
brainstorm
optional thinking
unsloth
conversational
8-bit precision
Update README.md
Browse files
README.md
CHANGED
|
@@ -40,7 +40,20 @@ pipeline_tag: text-generation
|
|
| 40 |
|
| 41 |
# Qwen3-Yoyo-V3-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-qx86x-hi-mlx
|
| 42 |
|
| 43 |
-
This
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 44 |
converted to MLX format from [DavidAU/Qwen3-Yoyo-V3-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG](https://huggingface.co/DavidAU/Qwen3-Yoyo-V3-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG)
|
| 45 |
using mlx-lm version **0.28.3**.
|
| 46 |
|
|
|
|
| 40 |
|
| 41 |
# Qwen3-Yoyo-V3-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-qx86x-hi-mlx
|
| 42 |
|
| 43 |
+
This is a new-old-stock version of the model, with embeddings at 8 bit.
|
| 44 |
+
|
| 45 |
+
The original [Qwen3-Yoyo-V3-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-qx86-hi-mlx](https://huggingface.co/nightmedia/Qwen3-Yoyo-V3-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-qx86-hi-mlx) is using 6 bit embeddings
|
| 46 |
+
|
| 47 |
+
```bash
|
| 48 |
+
Perplexity: 4.431 ± 0.031
|
| 49 |
+
Peak memory: 43.43 GB
|
| 50 |
+
```
|
| 51 |
+
|
| 52 |
+
Metrics coming soon. If this proves better than the qx86-hi, it will replace it in the catalog.
|
| 53 |
+
|
| 54 |
+
-G
|
| 55 |
+
|
| 56 |
+
This model [Qwen3-Yoyo-V3-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-qx86x-hi-mlx](https://huggingface.co/nightmedia/Qwen3-Yoyo-V3-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-qx86x-hi-mlx) was
|
| 57 |
converted to MLX format from [DavidAU/Qwen3-Yoyo-V3-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG](https://huggingface.co/DavidAU/Qwen3-Yoyo-V3-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG)
|
| 58 |
using mlx-lm version **0.28.3**.
|
| 59 |
|