Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

nvidia
/
Llama-3_3-Nemotron-Super-49B-v1_5

Text Generation
Transformers
Safetensors
PyTorch
English
nemotron-nas
nvidia
llama-3
conversational
custom_code
Model card Files Files and versions
xet
Community
12
New discussion
Resources
  • PR & discussions documentation
  • Code of Conduct
  • Hub documentation

variable_cache.py compatibility for v4.57.2 / python3.12

1
#12 opened 6 days ago by
NePe

cannot import name 'NEED_SETUP_CACHE_CLASSES_MAPPING'

1
#11 opened 7 days ago by
uygarkurt

Trying to fix issues with extra arguments to the model

#10 opened 18 days ago by
shmuli

Since `transformers` v4.56.0` the dictionary `ALL_STATIC_CACHE_IMPLEMENTATIONS` replaced `NEED_SETUP_CACHE_CLASSES_MAPPING`

4
#9 opened 22 days ago by
blewis-hir

Does vllm deployment supports --enable-reasoning and --reasoning-parser

#8 opened 2 months ago by
defactocorpse

Possible to disable thinking via a karg?

1
#7 opened 3 months ago by
SuperbEmphasis

Cannot disable thinking mode

1
#5 opened 4 months ago by
AekDevDev

Tool calling no stream

1
#4 opened 4 months ago by
yuchenxie

FP8 Quants please

3
#3 opened 4 months ago by
rjmehta

there will be a nemotron ultra v1_5?

❤️ 2
2
#2 opened 4 months ago by
bobox

Missing `modeling_decilm.py` when loading the model

3
#1 opened 4 months ago by
shawn2333
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs