json-structured / README.md
Dev8709's picture
Updated
44b5c36
---
title: Json Structured
emoji: πŸƒ
colorFrom: red
colorTo: gray
sdk: gradio
sdk_version: 5.33.0
app_file: app.py
pinned: false
short_description: Plain text to json using llama.cpp
---
# Plain Text to JSON with llama.cpp
This Hugging Face Space converts plain text into structured JSON format using llama.cpp for efficient CPU inference, powered by the Osmosis Structure 0.6B model.
## Features
- **llama.cpp Integration**: Uses llama-cpp-python for efficient CPU model inference
- **Osmosis Structure Model**: Specialized 0.6B parameter model for structured data extraction
- **Gradio Interface**: User-friendly web interface
- **JSON Conversion**: Converts unstructured text to well-formatted JSON
- **Auto-Download**: Automatically downloads the Osmosis model on first use
- **Demo Mode**: Basic functionality without requiring the AI model
## Setup
The space automatically installs:
- `llama-cpp-python` for llama.cpp integration
- Required build tools (`build-essential`, `cmake`)
- Gradio and other dependencies
- Downloads Osmosis Structure 0.6B model (~1.2GB) on first use
## Usage
1. **Quick Start**: Run `python setup_and_run.py` for automated setup
2. **Demo Mode**: Use "Demo (No Model)" for basic text-to-JSON conversion
3. **Full Mode**: Click "Load Model" to download and use the Osmosis model
4. **Customize**: Adjust temperature and max_tokens for different output styles
## Model Details
- **Model**: Osmosis Structure 0.6B BF16 GGUF
- **Repository**: https://huggingface.co/osmosis-ai/Osmosis-Structure-0.6B
- **Specialization**: Structure extraction and JSON generation
- **Size**: ~1.2GB download
- **Format**: GGUF (optimized for llama.cpp)
## Configuration
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference