Update README.md
Browse files
README.md
CHANGED
|
@@ -12,14 +12,15 @@ base_model:
|
|
| 12 |
|
| 13 |
Original Model : [THUDM/GLM-Z1-9B-0414](https://huggingface.co/THUDM/GLM-Z1-9B-0414)
|
| 14 |
|
| 15 |
-
Llama.cpp build:
|
| 16 |
|
| 17 |
I used imatrix to create all these quants using this [Dataset](https://gist.github.com/bartowski1182/eb213dccb3571f863da82e99418f81e8).
|
| 18 |
|
| 19 |
-
|
| 20 |
-
Update-
|
| 21 |
-
* [Fixed Quant] Re-quantized all quants
|
| 22 |
-
|
|
|
|
| 23 |
|
| 24 |
| | CPU (AVX2) | CPU (ARM NEON) | Metal | cuBLAS | rocBLAS | SYCL | CLBlast | Vulkan | Kompute |
|
| 25 |
| :------------ | :---------: | :------------: | :---: | :----: | :-----: | :---: | :------: | :----: | :------: |
|
|
|
|
| 12 |
|
| 13 |
Original Model : [THUDM/GLM-Z1-9B-0414](https://huggingface.co/THUDM/GLM-Z1-9B-0414)
|
| 14 |
|
| 15 |
+
Llama.cpp build: ced44be3 (5199)
|
| 16 |
|
| 17 |
I used imatrix to create all these quants using this [Dataset](https://gist.github.com/bartowski1182/eb213dccb3571f863da82e99418f81e8).
|
| 18 |
|
| 19 |
+
---
|
| 20 |
+
# Update-02
|
| 21 |
+
* [Fixed Quant] Re-quantized all quants to fix this issue. [#13099](https://github.com/ggml-org/llama.cpp/pull/13099) [#13140](https://github.com/ggml-org/llama.cpp/pull/13140)
|
| 22 |
+
|
| 23 |
+
---
|
| 24 |
|
| 25 |
| | CPU (AVX2) | CPU (ARM NEON) | Metal | cuBLAS | rocBLAS | SYCL | CLBlast | Vulkan | Kompute |
|
| 26 |
| :------------ | :---------: | :------------: | :---: | :----: | :-----: | :---: | :------: | :----: | :------: |
|