Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -4,7 +4,6 @@
|
|
4 |
# from langchain_huggingface import HuggingFaceEndpoint,HuggingFacePipeline, ChatHuggingFace
|
5 |
# from langchain_core.messages import HumanMessage, SystemMessage, AIMessage
|
6 |
|
7 |
-
# os.environ["HF_TOKEN"]=os.getenv('Ayush')
|
8 |
# os.environ["HUGGINGFACEHUB_API_KEY"]=os.getenv('Ayush')
|
9 |
|
10 |
# llama_model = HuggingFaceEndpoint(repo_id= "meta-llama/Llama-3.2-3B-Instruct",provider= "nebius",temperature=0.6, max_new_tokens=70,task="conversational")
|
@@ -23,6 +22,8 @@ from fpdf import FPDF
|
|
23 |
|
24 |
# Set HuggingFace token from env or st.secrets
|
25 |
os.environ["HUGGINGFACEHUB_API_TOKEN"] = os.getenv("keys")
|
|
|
|
|
26 |
|
27 |
# Topic-wise base prompts and models
|
28 |
topic_config = {
|
|
|
4 |
# from langchain_huggingface import HuggingFaceEndpoint,HuggingFacePipeline, ChatHuggingFace
|
5 |
# from langchain_core.messages import HumanMessage, SystemMessage, AIMessage
|
6 |
|
|
|
7 |
# os.environ["HUGGINGFACEHUB_API_KEY"]=os.getenv('Ayush')
|
8 |
|
9 |
# llama_model = HuggingFaceEndpoint(repo_id= "meta-llama/Llama-3.2-3B-Instruct",provider= "nebius",temperature=0.6, max_new_tokens=70,task="conversational")
|
|
|
22 |
|
23 |
# Set HuggingFace token from env or st.secrets
|
24 |
os.environ["HUGGINGFACEHUB_API_TOKEN"] = os.getenv("keys")
|
25 |
+
os.environ["HF_TOKEN"]=os.getenv('Ayush')
|
26 |
+
|
27 |
|
28 |
# Topic-wise base prompts and models
|
29 |
topic_config = {
|