Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -230,7 +230,7 @@ logging.basicConfig(level=logging.INFO)
|
|
230 |
logger = logging.getLogger(__name__)
|
231 |
|
232 |
# Load model and tokenizer
|
233 |
-
model_name = "FreedomIntelligence/Apollo-
|
234 |
# model_name = "emilyalsentzer/Bio_ClinicalBERT"
|
235 |
# model_name = "FreedomIntelligence/Apollo-2B"
|
236 |
|
@@ -239,7 +239,6 @@ model = AutoModelForCausalLM.from_pretrained(model_name)
|
|
239 |
|
240 |
tokenizer.pad_token = tokenizer.eos_token
|
241 |
|
242 |
-
|
243 |
app = FastAPI(title="Apollo RAG Medical Chatbot")
|
244 |
|
245 |
generation_config = GenerationConfig(
|
@@ -338,4 +337,4 @@ async def chat_fn(query: Query):
|
|
338 |
|
339 |
except Exception as e:
|
340 |
logger.error(f"Inference failed: {str(e)}")
|
341 |
-
raise HTTPException(status_code=500, detail="Model inference failed.")
|
|
|
230 |
logger = logging.getLogger(__name__)
|
231 |
|
232 |
# Load model and tokenizer
|
233 |
+
model_name = "FreedomIntelligence/Apollo-7B"
|
234 |
# model_name = "emilyalsentzer/Bio_ClinicalBERT"
|
235 |
# model_name = "FreedomIntelligence/Apollo-2B"
|
236 |
|
|
|
239 |
|
240 |
tokenizer.pad_token = tokenizer.eos_token
|
241 |
|
|
|
242 |
app = FastAPI(title="Apollo RAG Medical Chatbot")
|
243 |
|
244 |
generation_config = GenerationConfig(
|
|
|
337 |
|
338 |
except Exception as e:
|
339 |
logger.error(f"Inference failed: {str(e)}")
|
340 |
+
raise HTTPException(status_code=500, detail="Model inference TimeOut failed.")
|