import gradio as gr from PIL import Image import spaces from transformers import AutoModelForImageSegmentation import torch from torchvision import transforms import requests from io import BytesIO import os # --- Model and Processor Setup --- # Use a higher precision for matrix multiplication for better performance torch.set_float32_matmul_precision("high") # Load the BiRefNet model for image segmentation # trust_remote_code=True is required for this model birefnet = AutoModelForImageSegmentation.from_pretrained( "ZhengPeng7/BiRefNet", trust_remote_code=True ) # Move the model to the available device (GPU if available, otherwise CPU) device = "cuda" if torch.cuda.is_available() else "cpu" birefnet.to(device) # Define the image transformation pipeline transform_image = transforms.Compose( [ transforms.Resize((1024, 1024)), transforms.ToTensor(), transforms.Normalize([0.485, 0.456, 0.406], [0.229, 0.224, 0.225]), ] ) # --- Helper Function to Load Images --- def load_image(image_source, output_type="pil"): """ Loads an image from a file path, URL, or numpy array. """ if image_source is None: return None if isinstance(image_source, str): if image_source.startswith("http"): try: response = requests.get(image_source) response.raise_for_status() image = Image.open(BytesIO(response.content)) except requests.exceptions.RequestException as e: raise gr.Error(f"Could not fetch image from URL: {e}") else: image = Image.open(image_source) elif hasattr(image_source, 'shape'): # Check if it's a numpy-like array image = Image.fromarray(image_source) else: image = image_source # Assume it's already a PIL image if output_type == "pil": return image.convert("RGB") return image # --- Core Processing Function --- # Use @spaces.GPU decorator if you plan to run this on a GPU-enabled Hugging Face Space # @spaces.GPU def process_image_to_transparent(image: Image.Image) -> Image.Image: """ Takes a PIL image, removes the background, and returns a PIL image with an alpha channel. """ if image is None: return None image_size = image.size # Unsqueeze adds a batch dimension, which the model expects input_tensor = transform_image(image).unsqueeze(0).to(device) # Prediction without tracking gradients for efficiency with torch.no_grad(): # The model returns multiple outputs; the last one is the primary segmentation map preds = birefnet(input_tensor)[-1].sigmoid().cpu() # Process the prediction tensor to create a mask pred_tensor = preds[0].squeeze() mask_pil = transforms.ToPILImage()(pred_tensor) mask_resized = mask_pil.resize(image_size) # Apply the mask as an alpha channel to the original image image.putalpha(mask_resized) return image # --- Gradio Interface Functions --- def fn(image_source): """ Handles image uploads and URLs, returning the processed image. """ if image_source is None: return None pil_image = load_image(image_source, output_type="pil") processed_image = process_image_to_transparent(pil_image) return processed_image def process_file(image_filepath): """ Handles a single file upload and returns a downloadable processed file. """ if image_filepath is None: return None # Define the output path for the new PNG file base_name = os.path.basename(image_filepath.name) # Use .name for Gradio file objects name, _ = os.path.splitext(base_name) output_path = f"{name}_transparent.png" # Load the image from the provided file path pil_image = load_image(image_filepath.name, output_type="pil") # Process the image transparent_image = process_image_to_transparent(pil_image) # Save the processed image to the new path transparent_image.save(output_path) # Return the path to the newly created file for download return output_path # --- Gradio UI Definition --- # Define example images for the interface example_image_path = "butterfly.jpeg" # You should have a 'butterfly.jpeg' in the same directory or provide a full path # For demonstration, let's create a dummy example image if it doesn't exist. if not os.path.exists(example_image_path): print(f"'{example_image_path}' not found. Creating a dummy image for example.") try: dummy_img = Image.new('RGB', (200, 200), color = 'red') dummy_img.save(example_image_path) except Exception as e: print(f"Could not create dummy image: {e}") example_url = "https://i.ibb.co/67B6Knk9/students-1807505-1280.jpg" # Define the individual interfaces for each tab tab1 = gr.Interface( fn, inputs=gr.Image(label="Upload an Image", type="pil"), outputs=gr.Image(label="Processed Image", format="png"), examples=[[example_image_path]], api_name="image" ) tab2 = gr.Interface( fn, inputs=gr.Textbox(label="Paste an Image URL"), outputs=gr.Image(label="Processed Image", format="png"), examples=[[example_url]], api_name="text" ) tab3 = gr.Interface( process_file, inputs=gr.File(label="Upload an Image File"), outputs=gr.File(label="Download Processed PNG"), examples=[[example_image_path]], api_name="png" ) # Combine the interfaces into a tabbed layout demo = gr.TabbedInterface( [tab1, tab2, tab3], ["Image Upload", "URL Input", "File Output"], title="Background Removal Tool | CodeTechDevX" ) if __name__ == "__main__": demo.launch(show_error=True)