Man Cub
mancub
ยท
AI & ML interests
None yet
Recent Activity
new activity about 18 hours ago
LuffyTheFox/Qwen3.6-35B-A3B-Uncensored-Wasserstein-GGUF:OOM and context limits reached too soon new activity about 20 hours ago
rdtand/Qwen3.6-27B-PrismaQuant-5.5bit-vllm:Unable to run on 3090 new activity about 23 hours ago
AesSedai/Qwen3.6-35B-A3B-GGUF:Q6_K?Organizations
None yet
OOM and context limits reached too soon
1
#5 opened 1 day ago
by
mancub
Unable to run on 3090
1
#1 opened about 20 hours ago
by
mancub
How to split this model between 2 (3) GPUs and CPU/RAM ?
28
#12 opened about 1 month ago
by
mancub
My personal vLLM launch cmd on my old personal 2x3090 workstation
7
#1 opened about 2 months ago
by
tclf90
What was just updated and why?
๐ 1
2
#1 opened 19 days ago
by
mancub
How to use it with llama-server ?
๐ 1
3
#1 opened 30 days ago
by
mancub
Poor performance and pretty lobotomized
2
#1 opened about 1 month ago
by
mancub
Love the license, confused by some of the decisions.
๐ค๐ 16
15
#15 opened about 1 month ago
by
CyborgPaloma
It's really good.
๐ 1
26
#3 opened about 2 months ago
by
Shuasimodo
Increasing the precision of some of the weights when quantizing
๐ 4
57
#2 opened 2 months ago
by
Shuasimodo
A draft model with less parameters, for speculative thinking?
8
#5 opened 3 months ago
by
mancub
Jan 21: All GLM-4.7-Flash quants reuploaded - much better outputs!
๐ฅโค๏ธ 7
29
#10 opened 3 months ago
by
danielhanchen
Fast loras
2
#8 opened 4 months ago
by
melmass
Wan-Lighting : 4steps per model or 4steps total?
4
#59 opened 9 months ago
by
NielsGx
Can we have a Llama-3.1-8B-Lexi-Uncensored-V2_fp8_scaled.safetensors
๐ฅ 1
12
#10 opened about 1 year ago
by
drguolai
Within Seconds ?
7
#8 opened about 1 year ago
by
Daemontatox
Is it censored output?
12
#2 opened about 1 year ago
by
KurtcPhotoED
Please work with llama.cpp before releasing new models.
2
#10 opened about 1 year ago
by
bradhutchings
Lack of 33B models?
๐ 1
7
#1 opened over 2 years ago
by
mancub