ubergarm commited on
Commit
7621bae
·
1 Parent(s): 4af291b

Add smol-IQ5_KS

Browse files
Files changed (2) hide show
  1. README.md +57 -0
  2. images/perplexity.png +2 -2
README.md CHANGED
@@ -35,6 +35,8 @@ This first is just a "pure" test quant for baseline perplexity comparison:
35
  * `Q8_0` 664.295 GiB (8.504 BPW)
36
  - Final estimate: PPL = 3.3929 +/- 0.01985
37
 
 
 
38
  ## IQ5_K 464.062 GiB (5.941 BPW)
39
  Final estimate: PPL = 3.4000 +/- 0.01992
40
 
@@ -96,6 +98,61 @@ numactl -N 0 -m 0 \
96
 
97
  </details>
98
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
99
  ## IQ4_K 382.485 GiB (4.896 BPW)
100
  Final estimate: PPL = 3.4198 +/- 0.02009
101
 
 
35
  * `Q8_0` 664.295 GiB (8.504 BPW)
36
  - Final estimate: PPL = 3.3929 +/- 0.01985
37
 
38
+ *NOTE*: `smol` is convention indicating same size quantization for `ffn_(up|gate)_exps` and `ffn_down_exps` tensors.
39
+
40
  ## IQ5_K 464.062 GiB (5.941 BPW)
41
  Final estimate: PPL = 3.4000 +/- 0.01992
42
 
 
98
 
99
  </details>
100
 
101
+ ## smol-IQ5_KS 417.107 GiB (5.339 BPW)
102
+ Final estimate: PPL = 3.4059 +/- 0.01996
103
+
104
+ <details>
105
+
106
+ <summary>👈 Secret Recipe</summary>
107
+
108
+ ```bash
109
+ #!/usr/bin/env bash
110
+
111
+ custom="
112
+ ## Attention [0-60] (GPU)
113
+ blk\..*\.attn_k_b\.weight=q8_0
114
+ blk\..*\.attn_v_b\.weight=q8_0
115
+
116
+ # Balance of attn tensors
117
+ blk\..*\.attn_kv_a_mqa\.weight=q8_0
118
+ blk\..*\.attn_q_a\.weight=q8_0
119
+ blk\..*\.attn_q_b\.weight=q8_0
120
+ blk\..*\.attn_output\.weight=q8_0
121
+
122
+ ## First Three Dense Layers [0-2] (GPU)
123
+ blk\..*\.ffn_down\.weight=q8_0
124
+ blk\..*\.ffn_(gate|up)\.weight=q8_0
125
+
126
+ ## Shared Expert (1-60) (GPU)
127
+ blk\..*\.ffn_down_shexp\.weight=q8_0
128
+ blk\..*\.ffn_(gate|up)_shexp\.weight=q8_0
129
+
130
+ ## Routed Experts (1-60) (CPU)
131
+ blk\..*\.ffn_down_exps\.weight=iq5_ks
132
+ blk\..*\.ffn_(gate|up)_exps\.weight=iq5_ks
133
+
134
+ ## Token embedding and output tensors (GPU)
135
+ token_embd\.weight=iq6_k
136
+ output\.weight=iq6_k
137
+ "
138
+
139
+ custom=$(
140
+ echo "$custom" | grep -v '^#' | \
141
+ sed -Ez 's:\n+:,:g;s:,$::;s:^,::'
142
+ )
143
+
144
+ numactl -N 0 -m 0 \
145
+ ./build/bin/llama-quantize \
146
+ --custom-q "$custom" \
147
+ --imatrix /mnt/data/models/ubergarm/DeepSeek-V3.1-Terminus-GGUF/imatrix-DeepSeek-V3.1-Terminus-Q8_0.dat \
148
+ /mnt/data/models/ubergarm/DeepSeek-V3.1-Terminus-GGUF/DeepSeek-V3.1-Terminus-256x20B-safetensors-BF16-00001-of-00030.gguf \
149
+ /mnt/data/models/ubergarm/DeepSeek-V3.1-Terminus-GGUF/DeepSeek-V3.1-Terminus-smol-IQ5_KS.gguf \
150
+ IQ5_KS \
151
+ 192
152
+ ```
153
+
154
+ </details>
155
+
156
  ## IQ4_K 382.485 GiB (4.896 BPW)
157
  Final estimate: PPL = 3.4198 +/- 0.02009
158
 
images/perplexity.png CHANGED

Git LFS Details

  • SHA256: 5096fe5c326b907fc9357b77826c82f09cb940dcc878bd97f5fc5fcda49e54e6
  • Pointer size: 131 Bytes
  • Size of remote file: 163 kB

Git LFS Details

  • SHA256: 5311e77affc5a02d0dfaded8a472c517b5f0b62f2540340b5166e36ef9f32fdb
  • Pointer size: 131 Bytes
  • Size of remote file: 172 kB