nightmedia commited on
Commit
fa7c533
·
verified ·
1 Parent(s): f29ecc9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -0
README.md CHANGED
@@ -10,6 +10,31 @@ base_model: openai/gpt-oss-120b
10
 
11
  # gpt-oss-120b-qx64-mlx
12
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  This model [gpt-oss-120b-qx64-mlx](https://huggingface.co/gpt-oss-120b-qx64-mlx) was
14
  converted to MLX format from [openai/gpt-oss-120b](https://huggingface.co/openai/gpt-oss-120b)
15
  using mlx-lm version **0.27.1**.
 
10
 
11
  # gpt-oss-120b-qx64-mlx
12
 
13
+ The reason I created the qx64 and qx65 quants: I was looking to write some Perl as a Postgres function.
14
+
15
+ Most other quants simplify, offer really well written PL/PGSQL instead.
16
+
17
+ But I wanted PL/Perl. I am that guy.
18
+
19
+ The [qx65 quant](https://huggingface.co/nightmedia/gpt-oss-120b-qx65-mlx) has given me what I asked.
20
+
21
+ It followed instructions.
22
+
23
+ Then I asked the qx64 the same question--why did you follow my instructions: I showed it this prompt.
24
+
25
+ The qx65 gave me a very short, clean answer I could put as a comment in the code.
26
+
27
+ The qx64 gave me the history of PL/Perl and how many nice things I could do with it.
28
+
29
+ Until the performance metrics are available, please use these models with caution.
30
+
31
+ -G
32
+
33
+ ```bash
34
+ 75.26 tok/sec
35
+ 9338 tokens
36
+ 2.58s to first token
37
+ ```
38
  This model [gpt-oss-120b-qx64-mlx](https://huggingface.co/gpt-oss-120b-qx64-mlx) was
39
  converted to MLX format from [openai/gpt-oss-120b](https://huggingface.co/openai/gpt-oss-120b)
40
  using mlx-lm version **0.27.1**.