UVIS / README.md
DurgaDeepak's picture
Update README.md
ce1589d verified

A newer version of the Gradio SDK is available: 5.33.0

Upgrade
metadata
license: apache-2.0
title: UVIS
sdk: gradio
emoji: πŸ”₯
colorFrom: blue
colorTo: indigo
pinned: true
thumbnail: >-
  https://cdn-uploads.huggingface.co/production/uploads/6820d348853cd8d544c6b014/qapEjDg69wwVgeqCXWTiX.png
short_description: Unified Visual Intelligence System
allow_embedding: true

UVIS - Unified Visual Intelligence System

A Lightweight Web-Based Visual Perception Demo

Try it online: uvis.deecoded.io GitHub: github.com/DurgaDeepakValluri/UVIS


Overview

UVIS (Unified Visual Intelligence System) is a lightweight, web-based visual perception demo, originally conceptualized as a spin-off while building Perceptaβ€”a larger modular perception framework.

The goal of UVIS is to make scene understanding tools more accessible, allowing anyone to try object detection, semantic segmentation, and depth estimation through a clean web interface, without requiring local setup.

UVIS currently runs on Render.com's Free Tier, using lightweight models to ensure the experience remains stable on limited resources.


Key Features

Capability Description
🟒 Object Detection YOLOv5-Nano & YOLOv5-Small for fast, low-resource detection.
🟒 Semantic Segmentation SegFormer-B0 and DeepLabV3-ResNet50 for general-purpose scenes.
🟒 Depth Estimation MiDaS Small & DPT Lite for per-pixel depth estimation.
πŸ–ΌοΈ Scene Blueprint Unified overlay combining all selected tasks.
πŸ“Š Scene Metrics Scene complexity scoring and agent-friendly summaries.
πŸ“¦ Downloadable Results JSON, overlay images, and ZIP bundles.
🌐 Web-First Design No installation neededβ€”hosted live at uvis.deecoded.io.
πŸ› οΈ Open Source Contribution-friendly, easy to extend and improve.

Current Limitations & Roadmap

UVIS is designed for lightweight demos on free-tier hosting, which means:

  • Models are optimized for speed and minimal compute.
  • Only image input is supported at this time.

As the project grows and higher hosting tiers become available, the roadmap includes:

  • Video input support
  • Lightweight SLAM
  • Natural language scene descriptions
  • Higher-capacity, more accurate models

Architecture Highlights

  • Modular Python Backend with Model Registry
  • Streamlit-Based Interactive Web UI
  • HuggingFace Transformers & TorchVision Integration
  • Lightweight Model Support (Render-Compatible)
  • Structured JSON Output for AI Agents
  • Robust Error Handling and Logging

🀝 Contributing

UVIS is open-source and welcomes contributions. You can:

  • Suggest new features
  • Improve the web interface
  • Extend perception tasks
  • Report issues or bugs

πŸ’» Clone and Run Locally

git clone https://github.com/DurgaDeepakValluri/UVIS.git
cd UVIS
pip install -r requirements.txt

🌐 Live Demo

Explore it online at uvis.deecoded.io Upload an image, select your tasks, and view the resultsβ€”all in your browser.


πŸ“ License

Apache 2.0 License. Free for personal and commercial use with attribution. Β© 2025 Durga Deepak Valluri