cybrown commited on
Commit
28c236c
·
verified ·
1 Parent(s): 8ef1a52

Upload GGUF quantizations of heretic model

Browse files
.gitattributes CHANGED
@@ -33,3 +33,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ qwen3-4b-instruct-2507-heretic-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
37
+ qwen3-4b-instruct-2507-heretic-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
38
+ qwen3-4b-instruct-2507-heretic-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
39
+ qwen3-4b-instruct-2507-heretic-f16.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: p-e-w/Qwen3-4B-Instruct-2507-heretic
4
+ tags:
5
+ - gguf
6
+ - quantized
7
+ - heretic
8
+ - abliterated
9
+ - uncensored
10
+ model_type: qwen2
11
+ ---
12
+
13
+ # Qwen3-4B-Instruct-2507-heretic-GGUF
14
+
15
+ GGUF quantized versions of [p-e-w/Qwen3-4B-Instruct-2507-heretic](https://huggingface.co/p-e-w/Qwen3-4B-Instruct-2507-heretic) from "The Bestiary" collection.
16
+
17
+ ## Model Description
18
+
19
+ This is a GGUF conversion of the Qwen3-4B-Instruct-2507-heretic model, which is an abliterated (uncensored) version of Alibaba's Qwen2.5 4B Instruct model. The model has had its refusal mechanisms removed, making it more willing to engage with any prompt.
20
+
21
+ **Original Model:** [p-e-w/Qwen3-4B-Instruct-2507-heretic](https://huggingface.co/p-e-w/Qwen3-4B-Instruct-2507-heretic)
22
+ **Collection:** [The Bestiary by p-e-w](https://huggingface.co/collections/p-e-w/the-bestiary)
23
+
24
+ ## Quantization Formats
25
+
26
+ This repository contains 4 quantization levels:
27
+
28
+ | File | Size | Description | Use Case |
29
+ |------|------|-------------|----------|
30
+ | `qwen3-4b-instruct-2507-heretic-f16.gguf` | 7.5GB | Full 16-bit precision | Best quality, highest memory usage |
31
+ | `qwen3-4b-instruct-2507-heretic-Q8_0.gguf` | 4.0GB | 8-bit quantization | High quality, good balance |
32
+ | `qwen3-4b-instruct-2507-heretic-Q5_K_M.gguf` | 2.7GB | 5-bit quantization | Balanced quality/size |
33
+ | `qwen3-4b-instruct-2507-heretic-Q4_K_M.gguf` | 2.4GB | 4-bit quantization | Smallest size, good quality |
34
+
35
+ **Recommended:** `Q4_K_M` for most users (best balance of quality and size)
36
+
37
+ ## Usage
38
+
39
+ ### With Ollama
40
+
41
+ 1. Download the GGUF file you want to use
42
+ 2. Create a Modelfile:
43
+
44
+ ```
45
+ FROM ./qwen3-4b-instruct-2507-heretic-Q4_K_M.gguf
46
+
47
+ TEMPLATE """{{ if .System }}<|im_start|>system<|im_sep|>{{ .System }}<|im_end|>{{ end }}{{ if .Prompt }}<|im_start|>user<|im_sep|>{{ .Prompt }}<|im_end|>{{ end }}<|im_start|>assistant<|im_sep|>{{ .Response }}<|im_end|>"""
48
+
49
+ PARAMETER stop "<|im_start|>"
50
+ PARAMETER stop "<|im_end|>"
51
+ PARAMETER temperature 0.7
52
+ PARAMETER top_p 0.9
53
+ PARAMETER num_ctx 8192
54
+ ```
55
+
56
+ 3. Import to Ollama:
57
+ ```bash
58
+ ollama create qwen3-4b-heretic:Q4_K_M -f Modelfile
59
+ ```
60
+
61
+ 4. Run:
62
+ ```bash
63
+ ollama run qwen3-4b-heretic:Q4_K_M
64
+ ```
65
+
66
+ ### With llama.cpp
67
+
68
+ ```bash
69
+ ./llama-cli -m qwen3-4b-instruct-2507-heretic-Q4_K_M.gguf -p "Your prompt here" -n 512
70
+ ```
71
+
72
+ ### With Open WebUI
73
+
74
+ Once imported to Ollama, the model will automatically appear in the Open WebUI model dropdown.
75
+
76
+ ## Conversion Details
77
+
78
+ - **Converted using:** llama.cpp (latest)
79
+ - **Conversion date:** 2025-11-21
80
+ - **Base format:** FP16 GGUF
81
+ - **Quantization method:** llama-quantize
82
+
83
+ ## Important Note
84
+
85
+ This is an uncensored model with refusal mechanisms removed. Use responsibly and in accordance with applicable laws and regulations.
86
+
87
+ ## License
88
+
89
+ Inherits the Apache 2.0 license from the base Qwen model.
90
+
91
+ ## Credits
92
+
93
+ - **Original model:** Alibaba (Qwen2.5)
94
+ - **Abliteration:** p-e-w ([The Bestiary](https://huggingface.co/collections/p-e-w/the-bestiary))
95
+ - **GGUF conversion:** cybrown
qwen3-4b-instruct-2507-heretic-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf6aacb6e2bb0e7758555e8842c1bb57955a11aedeb7809ef16e3bf806a2306e
3
+ size 2497279104
qwen3-4b-instruct-2507-heretic-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67126e09d99ae1ea1921889175ef6e37834b1a160c5b53949ac7e1b349ba7615
3
+ size 2889512064
qwen3-4b-instruct-2507-heretic-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77f8f224eab65c11855b9144d2a09f437daa241df1214056249c1f919488b07b
3
+ size 4280403584
qwen3-4b-instruct-2507-heretic-f16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2ba00143246d0671e87d17be84a9af28cdfbbcbbf641307fe191f9c0c8cee82
3
+ size 8051283584