austinroy's picture
Update app.py
4392ec0 verified
# -*- coding: utf-8 -*-
"""Lookingsoft Radiology Assistant
Automatically generated by Colab.
This file is adapted from:
https://colab.research.google.com/#fileId=https%3A//storage.googleapis.com/kaggle-colab-exported-notebooks/rishirajacharya/i-o-25-radiology-with-medgemma-gemini-native-tts.b5cf5dca-3453-45b1-b7c0-ec7c22aedf1b.ipynb
# Lookingsoft Radiology Assistant
## Developed by Lookingsoft Team
This demo showcases an AI-powered radiology assistant that leverages MedGemma for medical image interpretation and Gemini’s native text-to-speech (TTS) for natural voice output. The assistant transforms complex radiology reports into easy-to-understand language and delivers it through a user-friendly voice-driven experience—highlighting key areas in radiology images and making insights more accessible.
### 🔐 Securing API Keys
We use secure tokens to authenticate with Hugging Face and Google’s Gemini APIs, ensuring safe and authorized access.
"""
import spaces
from google import genai
from google.genai import types
import os
gemini_api_key = os.getenv('GEMINI_API_KEY')
hf_token = os.environ.get("HF_TOKEN")
client = genai.Client(api_key=gemini_api_key)
"""### 🧠 Loading MedGemma for Radiology Insights
Here, we load the MedGemma model—an image-text model optimized for medical contexts. We apply 4-bit quantization to enhance performance and reduce memory usage on GPUs.
"""
from transformers import pipeline, BitsAndBytesConfig
import torch
model_kwargs = dict(torch_dtype=torch.bfloat16, device_map="cuda:0", quantization_config=BitsAndBytesConfig(load_in_4bit=True))
pipe = pipeline("image-text-to-text", model="google/medgemma-4b-it", model_kwargs=model_kwargs, token=hf_token)
pipe.model.generation_config.do_sample = False
"""### 🩻 Radiology Image Interpretation Logic
This function uses MedGemma to generate a plain-language report based on a given prompt and medical image. It formats the input and passes it to the model for inference.
"""
from PIL import Image
@spaces.GPU
def infer(prompt: str, image: Image.Image, system: str = None) -> str:
image_filename = "image.png"
image.save(image_filename)
messages = []
if system:
messages.append({
"role": "system",
"content": [{"type": "text", "text": system}]
})
messages.append({
"role": "user",
"content": [
{"type": "text", "text": prompt},
{"type": "image", "image": image}
]
})
output = pipe(text=messages, max_new_tokens=2048)
response = output[0]["generated_text"][-1]["content"]
return response
"""### 🔊 Prepare for Gemini's Native TTS
This helper function converts Gemini’s audio output into a `.wav` file—enabling the assistant to speak its reports in a natural-sounding voice.
"""
import wave
def wave_file(filename, pcm, channels=1, rate=24000, sample_width=2):
with wave.open(filename, "wb") as wf:
wf.setnchannels(channels)
wf.setsampwidth(sample_width)
wf.setframerate(rate)
wf.writeframes(pcm)
"""### 🤖 Integrating Image Analysis and Voice Output
This function combines the MedGemma analysis with Gemini’s TTS to produce both text and audio responses.
"""
import gradio as gr
import requests
def _do_predictions(text, image_file, image_url, source_type):
if source_type == "url":
image = Image.open(requests.get(image_url, headers={"User-Agent": "example"}, stream=True).raw)
else:
image = image_file
report = infer(text, image)
response = client.models.generate_content(
model="gemini-2.5-flash-preview-tts",
contents=report,
config=types.GenerateContentConfig(
response_modalities=["AUDIO"],
speech_config=types.SpeechConfig(
voice_config=types.VoiceConfig(
prebuilt_voice_config=types.PrebuiltVoiceConfig(
voice_name='Kore',
)
)
),
)
)
data = response.candidates[0].content.parts[0].inline_data.data
file_name='out.wav'
wave_file(file_name, data)
return report, file_name
"""### 🖼️ Interactive Web UI with Gradio
A user-friendly interface built with Gradio. Users can upload an image or provide a URL, enter a prompt, and receive both a text report and an audio explanation—powered by **MedGemma + Gemini TTS**.
"""
def toggle_image_src(choice):
if choice == "url":
return gr.update(visible=False), gr.update(visible=True)
else:
return gr.update(visible=True), gr.update(visible=False)
with gr.Blocks() as demo:
gr.Markdown(
"""
# Lookingsoft Radiology Assistant
## Developed by the Lookingsoft Team
This assistant demonstrates the integration of MedGemma for medical image interpretation with Gemini’s native text-to-speech (TTS). It simplifies complex radiology reports into clear, spoken language, making insights more accessible and understandable for both professionals and patients.
"""
)
with gr.Row():
with gr.Column():
with gr.Row():
text = gr.Text(label="Instructions", lines=2, interactive=True)
with gr.Column():
radio = gr.Radio(["file", "url"], value="file", label="Input Image Source")
image_file = gr.Image(label="File", type="pil", visible=True)
image_url = gr.Textbox(label="URL", visible=False)
with gr.Row():
submit = gr.Button("Generate")
with gr.Column():
output = gr.Textbox(label="Generated Report")
audio_output = gr.Audio(label="Generated Report (wav)")
submit.click(_do_predictions, inputs=[text, image_file, image_url, radio],
outputs=[output, audio_output])
radio.change(toggle_image_src, radio, [image_file, image_url], queue=False, show_progress=False)
gr.Examples(
fn=_do_predictions,
examples=[
["Describe this X-ray", Image.open(requests.get("https://google-rad-explain.hf.space/static/images/Effusion2.jpg", headers={"User-Agent": "example"}, stream=True).raw), None, "file"],
["Describe this CT", None, "https://google-rad-explain.hf.space/static/images/CT-Tumor.jpg", "url"],
],
inputs=[text, image_file, image_url, radio],
outputs=[output, audio_output]
)
gr.Markdown("""
### Disclaimer
This demonstration is for educational purposes only. It is not intended to diagnose or treat any disease or condition and should not be considered medical advice.
""")
demo.queue(max_size=8 * 4).launch(share=True)