inferoxy-hub / README.md
nazdridoy's picture
refactor(core): rebrand to AI-Inferoxy and update API endpoints
c2e6d7e verified
---
title: AI-Inferoxy AI Hub
emoji: 🚀
colorFrom: purple
colorTo: blue
sdk: gradio
app_file: app.py
pinned: false
hf_oauth: true
hf_oauth_authorized_org:
- nazdev
---
## 🚀 AI‑Inferoxy AI Hub
A focused, multi‑modal AI workspace. Chat, create images, transform images, generate short videos, and synthesize speech — all routed through AI‑Inferoxy for secure, quota‑aware token management and provider failover.
### Highlights
- Chat, Image, Image‑to‑Image, Video, and TTS in one app
- Works with any HF model exposed by your proxy
- Multi‑provider routing; default provider is `auto`
- Streaming chat and curated examples
### Quick Start (Hugging Face Space)
Add Space secrets:
- `PROXY_URL`: AI‑Inferoxy server URL (e.g., `https://proxy.example.com`)
- `PROXY_KEY`: API key for your proxy
Org access control: instead of a custom `ALLOWED_ORGS` secret and runtime checks, configure org restrictions in README metadata using `hf_oauth_authorized_org` per HF Spaces OAuth docs. Example:
```yaml
hf_oauth: true
hf_oauth_authorized_org:
- your-org-slug
- another-org
```
The app reads these at runtime — no extra setup required.
### How It Works
1. The app requests a valid token from AI‑Inferoxy for each call.
2. Requests are sent to the selected provider (or `auto`).
3. Status is reported back for rotation and telemetry.
### Using the App
- Chat: message → choose model/provider (`auto` by default) → tune temperature/top‑p/max tokens.
- Image: prompt → optional width/height (÷8), steps, guidance, seed, negative prompt.
- Image‑to‑Image: upload base image → describe the change → generate.
- Video: brief motion prompt → optional steps/guidance/seed.
- TTS: text → pick TTS model → adjust voice/style if supported.
### Configuration
- Model id only (e.g., `openai/gpt-oss-20b`, `stabilityai/stable-diffusion-xl-base-1.0`).
- Provider from dropdown. Default is `auto`.
### Providers
Compatible with providers configured in AI‑Inferoxy, including `auto` (default), `hf-inference`, `cerebras`, `cohere`, `groq`, `together`, `fal-ai`, `replicate`, `nebius`, `nscale`, and others.
### Security
- HF OAuth validates account; org membership is enforced by Space metadata (`hf_oauth_authorized_org`).
- Inference uses proxy‑managed tokens. Secrets are Space secrets.
- RBAC, rotation, and quarantine handled by AI‑Inferoxy.
### Troubleshooting
- 401/403: verify secrets and org access.
- 402/quota: handled by proxy; retry later or switch provider.
- Image size: width/height must be divisible by 8.
- Slow/failures: try smaller models, fewer steps, or another provider.
### License
This project is licensed under the GNU Affero General Public License v3.0 (AGPL-3.0) - see the [LICENSE](LICENSE) file for details.
### Links
- Live Space: [huggingface.co/spaces/nazdridoy/inferoxy-hub](https://huggingface.co/spaces/nazdridoy/inferoxy-hub)
- AI‑Inferoxy docs: [ai-inferoxy/huggingface-hub-integration](https://nazdridoy.github.io/ai-inferoxy/)
- Gradio docs: [gradio.app/docs](https://gradio.app/docs/)
— Built with AI‑Inferoxy for intelligent token management.