Spaces:
Runtime error
Runtime error
# installing required libraries in my_env | |
import requests | |
from PIL import Image | |
from transformers import AutoProcessor, BlipForConditionalGeneration | |
# Load the pretrained processor and model | |
processor = AutoProcessor.from_pretrained("Salesforce/blip-image-captioning-base") | |
model = BlipForConditionalGeneration.from_pretrained("Salesforce/blip-image-captioning-base") | |
# Load your image, DON'T FORGET TO WRITE YOUR IMAGE NAME | |
img_path = "demo.jpeg" | |
# convert it into an RGB format | |
image = Image.open(img_path).convert('RGB') | |
# You do not need a question for image captioning | |
text = "the image of" | |
inputs = processor(images=image, text=text, return_tensors="pt") | |
# Generate a caption for the image | |
outputs = model.generate(**inputs, max_length=50) | |
# Decode the generated tokens to text | |
caption = processor.decode(outputs[0], skip_special_tokens=True) | |
# Print the caption | |
print(caption) |