Melvin56 commited on
Commit
7c3d18d
·
verified ·
1 Parent(s): b69bcc1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -5
README.md CHANGED
@@ -12,14 +12,15 @@ base_model:
12
 
13
  Original Model : [THUDM/GLM-Z1-9B-0414](https://huggingface.co/THUDM/GLM-Z1-9B-0414)
14
 
15
- Llama.cpp build: 558a7647 (5190)
16
 
17
  I used imatrix to create all these quants using this [Dataset](https://gist.github.com/bartowski1182/eb213dccb3571f863da82e99418f81e8).
18
 
19
- ```
20
- Update-01
21
- * [Fixed Quant] Re-quantized all quants with build: 558a7647 (5190)
22
- ```
 
23
 
24
  | | CPU (AVX2) | CPU (ARM NEON) | Metal | cuBLAS | rocBLAS | SYCL | CLBlast | Vulkan | Kompute |
25
  | :------------ | :---------: | :------------: | :---: | :----: | :-----: | :---: | :------: | :----: | :------: |
 
12
 
13
  Original Model : [THUDM/GLM-Z1-9B-0414](https://huggingface.co/THUDM/GLM-Z1-9B-0414)
14
 
15
+ Llama.cpp build: ced44be3 (5199)
16
 
17
  I used imatrix to create all these quants using this [Dataset](https://gist.github.com/bartowski1182/eb213dccb3571f863da82e99418f81e8).
18
 
19
+ ---
20
+ # Update-02
21
+ * [Fixed Quant] Re-quantized all quants to fix this issue. [#13099](https://github.com/ggml-org/llama.cpp/pull/13099) [#13140](https://github.com/ggml-org/llama.cpp/pull/13140)
22
+
23
+ ---
24
 
25
  | | CPU (AVX2) | CPU (ARM NEON) | Metal | cuBLAS | rocBLAS | SYCL | CLBlast | Vulkan | Kompute |
26
  | :------------ | :---------: | :------------: | :---: | :----: | :-----: | :---: | :------: | :----: | :------: |