Issues with Tools use and Chat templates
#99 opened 12 months ago
by
pyrator
Upgrading Linux Dist
#98 opened 12 months ago
by
rkapuaala

Clone Repository
👍
2
1
#96 opened 12 months ago
by
clearcash

llama3.1 gguf format
3
#95 opened 12 months ago
by
davidomars
how can i use git clone Meta-Llama-3.1-8B-Instruct
2
#93 opened 12 months ago
by
xiangsuyu
Asking for Pro subscription
6
#92 opened about 1 year ago
by
Mayo133
update rope_scaling
#91 opened about 1 year ago
by
Arunjith
Update for correct tool use system prompt
👍
1
3
#90 opened about 1 year ago
by
ricklamers
What call() function parameters besides "query" can be used by the model when doing brave_search and wolfram_alpha tool calls?
#89 opened about 1 year ago
by
sszymczyk
What form of the built-in brave_search and wolfram_alpha tool call output is expected by the model?
➕
1
3
#88 opened about 1 year ago
by
sszymczyk
ValueError
1
#87 opened about 1 year ago
by
Bmurug3
Request: DOI
1
#86 opened about 1 year ago
by
sanjeev929
Request: DOI
1
#85 opened about 1 year ago
by
moh996
The model repeatedly outputs a large amount of text and does not comply with the instructs.
10
#84 opened about 1 year ago
by
baremetal
Llama repo access not aproved yet
#83 opened about 1 year ago
by
APaul1
Throwing Error for AutoModelForSequence Classification
1
#82 opened about 1 year ago
by
deshwalmahesh
GSM8K Evaluation Result: 84.5 vs. 76.95
17
#81 opened about 1 year ago
by
tanliboy

Deploying Llama3.1 to Nvidia T4 instance (sagemaker endpoints)
4
#80 opened about 1 year ago
by
mleiter
Variable answer is getting predicted for same prompt
#79 opened about 1 year ago
by
sjainlucky
Efficiency low after adding the adapter_model.safetensors with base model
#78 opened about 1 year ago
by
antony-pk

Minimum gpu ram capacity
🔥
2
12
#77 opened about 1 year ago
by
bob-sj
Tokenizer padding token
1
#76 opened about 1 year ago
by
Rish1
new tokenizer contains the cutoff date and today date by default
5
#74 opened about 1 year ago
by
yuchenlin

New bee questions
2
#73 opened about 1 year ago
by
rkapuaala

Add `base_model` metadata
#72 opened about 1 year ago
by
sbrandeis

Full SFT training caused lose its foundational capabilities
10
#71 opened about 1 year ago
by
sinlew
Wrong number of tensors; expected 292, got 291
6
#69 opened about 1 year ago
by
KingBadger
Fine tuned Meta-Llama-3.1-8B-Instruct deployment on AWS Sagemaker fails
➕
7
2
#68 opened about 1 year ago
by
byamasuwhatnowis

Quick Fix: Rope Scaling or Rope Type Error
4
#67 opened about 1 year ago
by
deepaksiloka
Can't reproduce MATH performance
1
#66 opened about 1 year ago
by
jpiabrantes
Banned for Iranian People
🚀
➕
9
15
#65 opened about 1 year ago
by
MustafaLotfi

Inference endpoint deployment for 'meta-llama/Meta-Llama-3.1-8B-Instruct' fails
➕
4
6
#62 opened about 1 year ago
by
Keertiraj
Meta-Llama-3.1-8B-Instruct deployment on AWS Sagemaker fails
3
#61 opened about 1 year ago
by
Keertiraj
Error Loading the original model file consolidated.00.pth from local
3
#60 opened about 1 year ago
by
chanduvkp

Unable to deploy Meta-Llama-3.1-8B-Instruct model on Sagemaker
3
#58 opened about 1 year ago
by
axs531622
CUDA out of memory on RTX A5000 inference.
6
#57 opened about 1 year ago
by
RoberyanL
Update README.md to reflect correct transformers version
#56 opened about 1 year ago
by
priyakhandelwal
Update README.md to reflect correct transformers version
#55 opened about 1 year ago
by
priyakhandelwal
NotImplementedError: Could not run 'aten::_local_scalar_dense' with arguments from the 'Meta' backend.
➕
2
3
#54 opened about 1 year ago
by
duccio84
Some of you might be interested in my 'silly' experiment.
🧠
1
2
#52 opened about 1 year ago
by
ZeroWw
Updated config.json
#51 opened about 1 year ago
by
WestM
🚀 LMDeploy support Llama3.1 and its Tool Calling. An example of calling "Wolfram Alpha" to perform complex mathematical calculations can be found from here!
#50 opened about 1 year ago
by
vansin

HF pro subscription for llama 3.1-8b
4
#49 opened about 1 year ago
by
ostoslista
Significant bias
👀
👍
3
6
#48 opened about 1 year ago
by
stutteringp0et
`rope_scaling` must be a dictionary with two fields
➕
🤝
8
4
#46 opened about 1 year ago
by
thunderdagger
Unable to load Llama 3.1 to Text-Genration WebUI
4
#45 opened about 1 year ago
by
keeeeesz
BUG Chat template doesn't respect `add_generation_prompt`flag from transformers tokenizer
👍
3
1
#44 opened about 1 year ago
by
ilu000

How to use the ASR on LLama3.1
👀
🔥
1
1
#43 opened about 1 year ago
by
andrygasy