Spaces:
No application file
No application file
Update app.py
Browse files
app.py
CHANGED
@@ -2,18 +2,20 @@ import gradio as gr
|
|
2 |
import numpy as np
|
3 |
import onnxruntime as ort
|
4 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
5 |
-
from huggingface_hub import hf_hub_download
|
6 |
import torch
|
|
|
|
|
|
|
7 |
|
8 |
HF_MODEL_ID = "mistralai/Mistral-Nemo-Instruct-2407"
|
9 |
HF_ONNX_REPO = "techAInewb/mistral-nemo-2407-fp32"
|
10 |
ONNX_MODEL_FILE = "model.onnx"
|
11 |
|
12 |
# Load tokenizer
|
13 |
-
tokenizer = AutoTokenizer.from_pretrained(HF_MODEL_ID)
|
14 |
|
15 |
# Load PyTorch model
|
16 |
-
pt_model = AutoModelForCausalLM.from_pretrained(HF_MODEL_ID, torch_dtype=torch.float32)
|
17 |
pt_model.eval()
|
18 |
|
19 |
# Load ONNX model
|
|
|
2 |
import numpy as np
|
3 |
import onnxruntime as ort
|
4 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
|
|
5 |
import torch
|
6 |
+
from huggingface_hub import hf_hub_download, HfFolder
|
7 |
+
|
8 |
+
token = HfFolder.get_token() or os.getenv("HF_TOKEN")
|
9 |
|
10 |
HF_MODEL_ID = "mistralai/Mistral-Nemo-Instruct-2407"
|
11 |
HF_ONNX_REPO = "techAInewb/mistral-nemo-2407-fp32"
|
12 |
ONNX_MODEL_FILE = "model.onnx"
|
13 |
|
14 |
# Load tokenizer
|
15 |
+
tokenizer = AutoTokenizer.from_pretrained(HF_MODEL_ID, token=token)
|
16 |
|
17 |
# Load PyTorch model
|
18 |
+
pt_model = AutoModelForCausalLM.from_pretrained(HF_MODEL_ID, torch_dtype=torch.float32, token=token)
|
19 |
pt_model.eval()
|
20 |
|
21 |
# Load ONNX model
|