Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -11,7 +11,7 @@ model_path = "meta-llama/Llama-2-7b-chat-hf"
|
|
11 |
tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False, use_auth_token=hf_token)
|
12 |
model = AutoModelForCausalLM.from_pretrained(model_path, use_auth_token=hf_token).half().cuda()
|
13 |
|
14 |
-
client = Client("https://fffiloni-test-llama-api-debug.hf.space/", hf_token=hf_token)
|
15 |
|
16 |
clipi_client = Client("https://fffiloni-clip-interrogator-2.hf.space/")
|
17 |
|
|
|
11 |
tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False, use_auth_token=hf_token)
|
12 |
model = AutoModelForCausalLM.from_pretrained(model_path, use_auth_token=hf_token).half().cuda()
|
13 |
|
14 |
+
#client = Client("https://fffiloni-test-llama-api-debug.hf.space/", hf_token=hf_token)
|
15 |
|
16 |
clipi_client = Client("https://fffiloni-clip-interrogator-2.hf.space/")
|
17 |
|