will
willfalco
ยท
AI & ML interests
None yet
Recent Activity
new activity 1 day ago
turboderp/Qwen3.5-35B-A3B-exl3:vision? liked a model 5 days ago
MetaphoricalCode/Qwen3.5-27B-heretic-v2-exl3-3bpw-hb6 liked a model 5 days ago
QuantTrio/Qwen3.5-122B-A10B-AWQOrganizations
None yet
sglang vllm?
1
#2 opened 5 days ago
by
willfalco
Anyone get this working on 4x RTX 6000 Pro?
๐ 2
3
#1 opened 26 days ago
by
zenmagnets
is it the same as other mixed ones?
1
#1 opened 17 days ago
by
willfalco
MMLU PRO Benchmark
3
#3 opened about 1 month ago
by
sevapru
6 x RTX 6000?
#1 opened 27 days ago
by
willfalco
Possible to run on six RTX Pro 6000 Blackwell with vLLM oder SGLang?
๐ 3
7
#2 opened about 1 month ago
by
FabianHeller
Great Model! - sglang mtp support for triton backend
๐ 3
4
#19 opened 3 months ago
by
chriswritescode
[request] DeepSeek-V3.1-Terminus
4
#3 opened 3 months ago
by
willfalco
you know which nightly it worked with? because it does not with current one
31
#1 opened 4 months ago
by
willfalco
random atrifacts on larger outputs
2
#4 opened 3 months ago
by
willfalco
is NVFP4 supported on sm120 (blackwell rtx pro 6000, rtx 5090 etc)?
10
#4 opened 4 months ago
by
Fernanda24
4 x RTX PRO 6000
๐ 1
2
#1 opened 4 months ago
by
willfalco
Is it possible to make smaller NVFP4 quant at 340-360GB to fit in 4x96gb?
๐ 1
68
#1 opened 4 months ago
by
Fernanda24
Question will it work in vllm or sglang with rtx 6000 blackwells? cuda arch sm120
6
#1 opened 5 months ago
by
Fernanda24
ooof this fits in 4x96gb can we get this for the new 3.2 Speciale ase well please :)
16
#2 opened 4 months ago
by
Fernanda24
Aww Man!
20
#1 opened 4 months ago
by
mtcl
anyone ran this on blackwell?
๐ฅ 1
#2 opened 4 months ago
by
willfalco
you know which nightly it worked with? because it does not with current one
31
#1 opened 4 months ago
by
willfalco