Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -268,6 +268,8 @@ generation_config = GenerationConfig(
|
|
268 |
pipe = TextGenerationPipeline(
|
269 |
model=model,
|
270 |
tokenizer=tokenizer,
|
|
|
|
|
271 |
device=model.device.index if torch.cuda.is_available() else -1
|
272 |
)
|
273 |
|
@@ -344,7 +346,8 @@ async def chat_fn(query: Query):
|
|
344 |
@app.post("/ask-rag")
|
345 |
async def chat_fn(query: Query):
|
346 |
message = query.message
|
347 |
-
|
|
|
348 |
|
349 |
# Run RAG inference in thread
|
350 |
loop = asyncio.get_event_loop()
|
|
|
268 |
pipe = TextGenerationPipeline(
|
269 |
model=model,
|
270 |
tokenizer=tokenizer,
|
271 |
+
generation_config = generation_config,
|
272 |
+
task = "text-generation",
|
273 |
device=model.device.index if torch.cuda.is_available() else -1
|
274 |
)
|
275 |
|
|
|
346 |
@app.post("/ask-rag")
|
347 |
async def chat_fn(query: Query):
|
348 |
message = query.message
|
349 |
+
prompt = generate_prompt(message)
|
350 |
+
logger.info(f"Received message: {prompt}")
|
351 |
|
352 |
# Run RAG inference in thread
|
353 |
loop = asyncio.get_event_loop()
|