# Deploying GAIA on Hugging Face Spaces This guide provides step-by-step instructions for deploying the GAIA agent on Hugging Face Spaces, making it accessible as a web application to users worldwide. ## Why Deploy on Hugging Face Spaces? Hugging Face Spaces offers several advantages for deploying GAIA: 1. **Free Hosting**: Basic deployment is free with reasonable usage limits 2. **Easy Sharing**: Public URL that can be shared with anyone 3. **Version Control**: Built-in Git integration 4. **Secrets Management**: Secure storage for API keys 5. **Community**: Integration with the broader AI community 6. **Customization**: Support for custom domains and branding ## Prerequisites Before deploying to Hugging Face Spaces, you'll need: 1. **Hugging Face Account**: Create an account at [huggingface.co](https://huggingface.co/join) 2. **API Keys**: Gather all necessary API keys (OpenAI, Serper, etc.) 3. **GAIA Repository**: A local copy of the GAIA repository 4. **Git**: For pushing your code to Hugging Face ## Deployment Steps ### Step 1: Prepare Your Repository 1. Clone the GAIA repository if you haven't already: ```bash git clone https://github.com/your-organization/gaia.git cd gaia ``` 2. Create a specific `requirements.txt` file for Hugging Face deployment: ```bash # Copy the main requirements cp requirements.txt requirements-hf.txt # Edit to add Gradio (if not already included) echo "gradio>=4.0.0" >> requirements-hf.txt # Remove any development-specific packages # Edit requirements-hf.txt to remove unnecessary packages ``` 3. Create a Hugging Face-specific `app.py` or modify the existing one: ```python import os import gradio as gr from src.gaia.agent import GaiaAgent from src.gaia.config import Configuration # Initialize configuration config = Configuration() # Set default values for Hugging Face deployment config.set("demo_mode", True) config.set("models.default", "gpt-3.5-turbo") # Use cheaper model by default # Initialize the agent agent = GaiaAgent(config=config) # Define the Gradio interface def process_query(query, history): try: response = agent.run(query) return response except Exception as e: return f"Error: {str(e)}" # Create the gradio app demo = gr.ChatInterface( fn=process_query, title="GAIA - Grounded AI Alignment Agent", description="Ask any question and GAIA will search for information to provide a grounded answer.", examples=[ "What is quantum computing?", "Explain the theory of relativity in simple terms.", "What are the latest developments in AI safety?" ], theme="huggingface" ) # Launch the app if __name__ == "__main__": demo.launch() ``` ### Step 2: Create a Space on Hugging Face 1. Log in to Hugging Face and go to [huggingface.co/spaces](https://huggingface.co/spaces) 2. Click on "Create new Space" 3. Fill in the details: - **Owner**: Your username or organization - **Space name**: Choose a unique name (e.g., "gaia-agent") - **License**: Choose an appropriate license (e.g., MIT) - **SDK**: Choose "Gradio" - **Space hardware**: Start with "CPU basic" (free tier) - **Make this Space private**: Optional, if you want to restrict access 4. Click "Create Space" ### Step 3: Configure the Repository 1. In your local GAIA directory, add the Hugging Face Space as a remote: ```bash git remote add space https://huggingface.co/spaces/your-username/gaia-agent ``` 2. Create a `.gitignore` file to exclude unnecessary files: ```bash cat > .gitignore << EOF # Python __pycache__/ *.py[cod] *$py.class *.so .Python build/ develop-eggs/ dist/ downloads/ eggs/ .eggs/ lib/ lib64/ parts/ sdist/ var/ wheels/ *.egg-info/ .installed.cfg *.egg # Virtual Environment venv/ ENV/ env/ # Local configuration .env config.json # Logs logs/ *.log # Results results/ # IDE files .idea/ .vscode/ *.swp *.swo EOF ``` 3. Create a `README.md` file for your Space: ```bash cat > README.md << EOF # GAIA Agent GAIA (Grounded AI Alignment) is an AI agent designed to answer questions with grounded, factual information. ## Features - Web search capabilities using multiple providers - Academic research integration - Reasoning tools for complex problems - Memory system for context-aware responses ## Usage Simply type your question in the input box and click "Submit" to get a response from GAIA. ## Examples - "What is quantum computing?" - "Explain the theory of relativity in simple terms." - "What are the latest developments in AI safety?" ## About GAIA is built with Python using LangChain, LangGraph, and GPT-4. It leverages various APIs to provide accurate and up-to-date information. EOF ``` ### Step 4: Create a Requirements File for Hugging Face Create or modify the `requirements.txt` file to include only the necessary packages: ```bash cat > requirements.txt << EOF gradio>=4.0.0 langchain>=0.0.267 langgraph>=0.0.15 openai>=1.1.1 tiktoken>=0.5.1 supabase>=2.0.3 requests>=2.31.0 python-dotenv>=1.0.0 EOF ``` ### Step 5: Set Up Environment Variables In your Hugging Face Space: 1. Go to the Settings tab of your Space 2. Scroll down to the "Repository secrets" section 3. Add your API keys and configuration as secrets: - `OPENAI_API_KEY`: Your OpenAI API key - `SERPER_API_KEY`: Your Serper API key (if used) - `PERPLEXITY_API_KEY`: Your Perplexity API key (if used) - Any other necessary API keys or configuration values ### Step 6: Push Your Code to Hugging Face Commit and push your code to the Hugging Face Space: ```bash # Add your files git add app.py requirements.txt README.md .gitignore src/ # Commit the changes git commit -m "Initial GAIA deployment on Hugging Face" # Push to Hugging Face git push space main ``` After pushing, Hugging Face will automatically build and deploy your application. This may take a few minutes. ### Step 7: Test Your Deployment 1. Once the build is complete, navigate to your Space's URL: `https://huggingface.co/spaces/your-username/gaia-agent` 2. Test the application by asking a few questions 3. Check the Space logs for any errors or issues: - Go to the "Settings" tab - Scroll down to "Factory reboot" section - Click on "View logs" ## Advanced Configuration ### Custom Domain To use a custom domain with your Space: 1. Go to the "Settings" tab of your Space 2. Scroll down to the "Custom domain" section 3. Enter your domain name (e.g., `gaia.yourdomain.com`) 4. Follow the instructions to set up DNS records ### Upgrading Hardware For better performance or to handle more traffic: 1. Go to the "Settings" tab of your Space 2. Scroll down to the "Space hardware" section 3. Choose a higher tier (note: this will incur costs) - CPU Upgrade: For faster processing - GPU: For model hosting or intensive processing - Memory Boost: For handling larger datasets ### Persistent Storage To enable persistent storage for your Space: 1. Go to the "Settings" tab of your Space 2. Scroll down to the "Persistent storage" section 3. Enable persistent storage (up to 10GB for free) This allows you to store data that persists between restarts, such as logs or cached results. ### Authentication To restrict access to authenticated users: 1. In your `app.py`, modify the launch parameters: ```python demo.launch(auth=("username", "password")) ``` 2. Or for more flexible authentication: ```python demo.launch(auth_message="Enter GAIA password", auth=lambda u, p: p == os.environ.get("GAIA_PASSWORD", "default_password")) ``` 3. Add the `GAIA_PASSWORD` environment variable in your Space settings ### Scheduled Restarts For long-running deployments, scheduled restarts can help maintain stability: 1. Go to the "Settings" tab of your Space 2. Scroll down to the "Factory reboot" section 3. Set a schedule for automatic reboots (e.g., daily or weekly) ## Optimizing for Hugging Face Spaces ### Reducing Startup Time To reduce startup time and improve user experience: 1. Lazy-load components when possible: ```python def load_agent(): if not hasattr(load_agent, "agent"): load_agent.agent = GaiaAgent(config=config) return load_agent.agent def process_query(query, history): agent = load_agent() return agent.run(query) ``` 2. Use caching for expensive operations: ```python import functools @functools.lru_cache(maxsize=100) def get_cached_result(query): # Expensive operation return result ``` ### Memory Usage Optimization To stay within Hugging Face's memory limits: 1. Limit model usage: ```python config.set("models.default", "gpt-3.5-turbo") # Uses less memory than GPT-4 ``` 2. Implement efficient memory management: ```python # Clear working memory after each query def process_query(query, history): agent = load_agent() result = agent.run(query) agent.reset(clear_memory=False) # Clear working memory but keep conversation history return result ``` 3. Disable memory-intensive features in the configuration: ```python config.set("memory.supabase.enabled", False) # Use simpler memory system config.set("tools.image_analysis.enabled", False) # Disable memory-intensive tools ``` ## Monitoring and Maintenance ### Monitoring Usage Monitor your Space's usage and performance: 1. Go to the "Settings" tab of your Space 2. Scroll down to the "Metrics" section 3. View CPU, memory, and disk usage over time ### Updating Your Deployment To update your GAIA deployment: 1. Make changes to your local repository 2. Commit and push to your Hugging Face Space: ```bash git add . git commit -m "Update GAIA deployment" git push space main ``` ### Handling Errors If you encounter errors in your deployment: 1. Check the Space logs for error messages 2. Implement better error handling in your code: ```python def process_query(query, history): try: agent = load_agent() return agent.run(query) except Exception as e: # Log the error print(f"Error processing query: {str(e)}") # Return a user-friendly message return "I'm having trouble processing your request. Please try again or ask a different question." ``` 3. Set up monitoring to be notified of errors ## Conclusion You now have GAIA deployed on Hugging Face Spaces, making it accessible as a web application. This allows you to share your agent with others, collaborate with the community, and benefit from Hugging Face's infrastructure. For more advanced deployments or custom integrations, consider exploring: - [Local Deployment Guide](local.md) for self-hosting options - [API Documentation](../api/agent.md) for programmatic integrations - [Hugging Face Spaces Documentation](https://huggingface.co/docs/hub/spaces) for more Spaces features If you encounter any issues or have questions, refer to the [troubleshooting section](#handling-errors) or create an issue on the project's GitHub repository.