-
-
-
-
-
-
Inference Providers
Active filters:
gptq
TheBloke/MythoMax-L2-13B-GPTQ
Text Generation
•
2B
•
Updated
•
1.34k
•
215
Qwen/Qwen3-30B-A3B-GPTQ-Int4
Text Generation
•
5B
•
Updated
•
60.9k
•
36
MidnightPhreaker/GLM-4.5-Air-REAP-82B-A12B-GPTQ-INT4-gs32
14B
•
Updated
•
17
•
2
TheBloke/Wizard-Vicuna-30B-Uncensored-GPTQ
Text Generation
•
4B
•
Updated
•
561
•
586
TheBloke/Psyfighter-13B-GPTQ
Text Generation
•
2B
•
Updated
•
13
•
7
Qwen/Qwen1.5-110B-Chat-GPTQ-Int4
Text Generation
•
17B
•
Updated
•
10.8k
•
18
Qwen/Qwen2-72B-Instruct-GPTQ-Int4
Text Generation
•
12B
•
Updated
•
3.43k
•
33
Qwen/Qwen2-1.5B-Instruct-GPTQ-Int4
Text Generation
•
0.6B
•
Updated
•
10.8k
•
5
Qwen/Qwen2-VL-7B-Instruct-GPTQ-Int4
Image-Text-to-Text
•
3B
•
Updated
•
12.8k
•
37
Qwen/Qwen2-VL-2B-Instruct-GPTQ-Int4
Image-Text-to-Text
•
1B
•
Updated
•
2.61k
•
26
Qwen/Qwen2.5-1.5B-Instruct-GPTQ-Int4
Text Generation
•
0.4B
•
Updated
•
3.68k
•
2
Qwen/Qwen2.5-7B-Instruct-GPTQ-Int4
Text Generation
•
2B
•
Updated
•
17.3k
•
29
Qwen/Qwen2.5-14B-Instruct-GPTQ-Int4
Text Generation
•
3B
•
Updated
•
20.4k
•
24
Qwen/Qwen2.5-32B-Instruct-GPTQ-Int8
Text Generation
•
10B
•
Updated
•
38k
•
12
Qwen/Qwen2.5-72B-Instruct-GPTQ-Int4
Text Generation
•
12B
•
Updated
•
17.5k
•
39
Qwen/Qwen2.5-Coder-7B-Instruct-GPTQ-Int4
Text Generation
•
2B
•
Updated
•
7.33k
•
9
shuyuej/Llama-3.2-1B-GPTQ
0.4B
•
Updated
•
171
•
1
Qwen/Qwen2.5-Coder-14B-Instruct-GPTQ-Int4
Text Generation
•
3B
•
Updated
•
18.7k
•
6
Qwen/Qwen2.5-Coder-32B-Instruct-GPTQ-Int4
Text Generation
•
6B
•
Updated
•
74.3k
•
20
Satwik11/Microsoft-phi-4-Instruct-AutoRound-GPTQ-4bit
3B
•
Updated
•
48
•
2
FunAGI/Qwen2.5-Omni-7B-GPTQ-4bit
Any-to-Any
•
5B
•
Updated
•
73
•
55
Qwen/Qwen3-1.7B-GPTQ-Int8
Text Generation
•
0.7B
•
Updated
•
3.4k
•
5
Qwen/Qwen3-235B-A22B-GPTQ-Int4
Text Generation
•
Updated
•
41.4k
•
24
Qwen/Qwen2.5-Omni-7B-GPTQ-Int4
Any-to-Any
•
5B
•
Updated
•
1.21k
•
11
QuantTrio/Qwen3-30B-A3B-Instruct-2507-GPTQ-Int8
Text Generation
•
8B
•
Updated
•
188k
•
8
QuantTrio/Qwen3-Coder-30B-A3B-Instruct-GPTQ-Int8
Text Generation
•
8B
•
Updated
•
1.37k
•
4
baichuan-inc/Baichuan-M2-32B-GPTQ-Int4
Text Generation
•
33B
•
Updated
•
713
•
11
jart25/Qwen3-Next-80B-A3B-Instruct-Int4-GPTQ
Updated
•
1.22k
•
2
jart25/Qwen3-Next-80B-A3B-Thinking-Int4-GPTQ
Updated
•
234
•
1
QuantTrio/Kimi-Dev-72B-GPTQ-Int4
Text Generation
•
73B
•
Updated
•
20
•
1