Fork of daryl149/llama-2-70b-chat-hf with max_position_embeddings
set to 4096 instead of 2048.
There are a few small differences between this and Meta's version
temperature
andtop_p
are specified in Meta's, but not here.- Meta uses
transformers
4.31.0.dev0 instead of 4.31.0.
- Downloads last month
- 3
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support