Spaces:
Sleeping
Sleeping
I need a DevOps/Infra repository with: | |
Tech Stack: | |
- Containerization: Docker & Docker Compose | |
- Orchestration: Hugging Face Spaces (Docker build) | |
- Database: Supabase (Postgres with SSL required) | |
- Integrations: GitHub, Google Cloud CLI, Vertex AI, LangChain, Community Nodes | |
- AI Features: Vector Store, AI Agents, AI Assistant, LangChain pipelines | |
- Automation: n8n (self-hosted Space), CI/CD with GitHub Actions | |
Core Features: | |
- Infrastructure as code for n8n running on Hugging Face Spaces | |
- Supabase Postgres as database backend (SSL enforced) | |
- Secure secrets injection (HF Space → environment variables) | |
- CI/CD pipeline that updates Space from GitHub repo | |
- Sync workflows via n8n API and export/backup into GitHub | |
- Integration with multiple GitHub repos as Knowledge Base: | |
- https://github.com/danilonovaisv/CHATGPT-knowledge-base/tree/main/projects/n8n | |
- https://github.com/danilonovaisv/CHATGPT-knowledge-base/tree/main/projects/videos-e-animacoes | |
- https://github.com/danilonovaisv/CHATGPT-knowledge-base/tree/main/projects/midjorney-prompt | |
- Vector database integration to store knowledge embeddings for workflows | |
- Built-in nodes + community nodes for LangChain, Google APIs, Vertex AI | |
Start with the following repository structure: | |
/n8n-infra | |
/docker | |
Dockerfile (pin n8n version, ENV configs) | |
docker-compose.yml (services: n8n, supabase connection, vector store) | |
/config | |
.env.example (template for secrets: DB_HOST, DB_USER, DB_PASS, ENCRYPTION_KEY, JWT_SECRET, WEBHOOK_URL, etc.) | |
credentials/ (to store API keys and OAuth tokens) | |
/workflows | |
backup/ (exported workflows JSON) | |
/knowledge | |
n8n/ (mirror from projects/n8n) | |
videos-e-animacoes/ (mirror from repo) | |
midjourney-prompt/ (mirror from repo) | |
/scripts | |
backup.sh (pg_dump for Supabase, workflow export) | |
restore.sh (restore db) | |
sync-knowledge.sh (pull repos, upsert into DB/vector store) | |
/github | |
workflows/ | |
deploy-to-hf.yml (GitHub Action: build & push image to HF Space) | |
backup-workflows.yml (GitHub Action: export workflows via API nightly) | |
sync-knowledge.yml (GitHub Action: pull knowledge repos and update DB/vector store) | |
README.md (instructions: how to deploy, secrets, CI/CD, rollback strategy) | |
Configuration & Integrations: | |
- Dockerfile pinned to a specific n8n version (no `latest`) | |
- Docker Compose with service for n8n + supabase connector + vector DB (like pgvector or chroma) | |
- GitHub Actions for CI/CD: | |
- build image | |
- push to Hugging Face Space | |
- trigger rebuild | |
- Backup workflow: | |
- pg_dump from Supabase | |
- export workflows from n8n API | |
- Sync job: | |
- clone repos from GitHub knowledge base | |
- ingest JSON/MD into vector DB | |
- Integrate LangChain + AI Assistant via community nodes | |
- Configure built-in nodes for Google/Vertex AI | |
- Add CLI tools for Google Cloud | |
Please leave placeholders for variables/secrets in .env.example: | |
- N8N_ENCRYPTION_KEY= | |
- N8N_USER_MANAGEMENT_JWT_SECRET= | |
- DB_TYPE=postgresdb | |
- DB_POSTGRESDB_HOST= | |
- DB_POSTGRESDB_PORT=5432 | |
- DB_POSTGRESDB_DATABASE= | |
- DB_POSTGRESDB_USER= | |
- DB_POSTGRESDB_PASSWORD= | |
- DB_POSTGRESDB_SSL=true | |
- WEBHOOK_URL= | |
- HF_TOKEN= | |
- GITHUB_TOKEN= | |
- GOOGLE_PROJECT_ID= | |
- GOOGLE_CREDENTIALS_PATH= | |
Document all steps in README with: | |
- How to deploy locally with Docker Compose | |
- How to deploy on Hugging Face Spaces | |
- How to configure Supabase | |
- How to run backups & restores | |
- How to integrate workflows with LangChain/Agents | |
- How to keep knowledge repos synced |