Spaces:
Paused
Paused
| import time | |
| from fastapi import FastAPI, Request, HTTPException | |
| from pydantic import BaseModel | |
| from g4f.client import Client | |
| import uvicorn | |
| app = FastAPI() | |
| API_PREFIX = "/" | |
| # Middleware for logging request time | |
| async def log_process_time(request: Request, call_next): | |
| start_time = time.time() | |
| response = await call_next(request) | |
| process_time = time.time() - start_time | |
| print(f"{request.method} {response.status_code} {request.url.path} {process_time*1000:.2f} ms") | |
| return response | |
| # Request body model | |
| class ChatCompletionRequest(BaseModel): | |
| model: str | |
| messages: list[dict] | |
| async def root(): | |
| return {"message": "API server running"} | |
| async def ping(): | |
| return {"message": "pong"} | |
| async def get_models(): | |
| return { | |
| "object": "list", | |
| "data": [ | |
| {"id": "gpt-4o-mini", "object": "model", "owned_by": "ddg"}, | |
| {"id": "claude-3-haiku", "object": "model", "owned_by": "ddg"}, | |
| {"id": "llama-3.1-70b", "object": "model", "owned_by": "ddg"}, | |
| {"id": "mixtral-8x7b", "object": "model", "owned_by": "ddg"}, | |
| {"id": "o3-mini", "object": "model", "owned_by": "ddg"}, | |
| ], | |
| } | |
| async def chat_completions(request: ChatCompletionRequest): | |
| try: | |
| # Only using DuckAI directly | |
| content = " ".join([msg.get("content", "") for msg in request.messages]) | |
| client = Client() | |
| response = client.chat.completions.create( | |
| model=request.model, | |
| messages=[{"role": "user", "content": content}], | |
| web_search=False | |
| ) | |
| return response | |
| except Exception as e: | |
| raise HTTPException(status_code=500, detail=str(e)) | |
| if __name__ == "__main__": | |
| uvicorn.run("app:app", host="0.0.0.0", port=7860, reload=True) |