Kkk
Michalea
ยท
AI & ML interests
None yet
Recent Activity
new activity 2 days ago
BLR2/Qwen3.5-9B-Eagle3-ShareGPT:MTP vs Eagle3 new activity 4 days ago
nvidia/NVIDIA-Nemotron-3-Super-120B-A12B-BF16:Add SWE-Bench Verified evaluation resultsOrganizations
None yet
MTP vs Eagle3
2
#1 opened 14 days ago
by
Michalea
Add SWE-Bench Verified evaluation results
2
#18 opened 12 days ago
by
nielsr
SciBench Results - methodology
#3 opened 4 days ago
by
Michalea
Eval. of your quant.
#5 opened 7 days ago
by
Michalea
math-oai.yaml file for aime eval
#56 opened 7 days ago
by
Michalea
SGLang and MTP
1
#2 opened 22 days ago
by
Michalea
RoPE theta 5mln instead of 1mln
#15 opened about 1 month ago
by
Michalea
New activity in lmsys/SGLang-EAGLE3-Qwen3-Next-80B-A3B-Instruct-FP8-SpecForge-Meituan about 1 month ago
The comparison with the original MTP
๐ 1
1
#2 opened about 1 month ago
by
Michalea
context length/number of generated tokens during training
#4 opened about 1 month ago
by
Michalea
Datasets used to create this head
#5 opened about 2 months ago
by
Michalea
A low number of evaluation benchmarks
2
#5 opened about 2 months ago
by
Michalea
Context length and regeneration
4
#1 opened about 2 months ago
by
Michalea
MTP quality, 47 layer
3
#7 opened about 2 months ago
by
Michalea
Efficiency of NVFP4 vs FP16/8
โ 3
#4 opened about 2 months ago
by
Michalea
opis wersji 3.0
2
#1 opened 3 months ago
by
jacek2024
Evaluation
๐ 4
#3 opened about 2 months ago
by
Michalea
Severe Looping/Repetitive Output when using --kv-cache-dtype fp8 with GLM-4.7-Flash-FP8-Dynamic on vLLM
4
#2 opened about 2 months ago
by
ShelterW
Inconsistent description with the evaluation results
1
#3 opened 3 months ago
by
Michalea