File size: 1,076 Bytes
da78f41 b201b46 da78f41 b201b46 da78f41 b201b46 da78f41 b201b46 da78f41 b201b46 da78f41 b201b46 da78f41 b201b46 da78f41 b201b46 da78f41 b201b46 da78f41 b201b46 da78f41 b201b46 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 |
---
tags:
- clip
- vision-language
- image-text
- pytorch
license: apache-2.0
---
# CLIP Model
This is a fine-tuned CLIP model for vision-language tasks.
## Model Description
This model was fine-tuned from a base CLIP model and includes custom temperature scaling.
## Usage
```python
from transformers import CLIPModel, CLIPProcessor
import torch
# Load model and processor
model = CLIPModel.from_pretrained("aprendesc/CLIP_model_v0")
processor = CLIPProcessor.from_pretrained("aprendesc/CLIP_model_v0")
# Load temperature parameter if available
try:
from huggingface_hub import hf_hub_download
temperature_path = hf_hub_download(repo_id="aprendesc/CLIP_model_v0", filename="temperature.pth")
temperature = torch.load(temperature_path, map_location='cpu')
print(f"Temperature parameter: {temperature}")
except:
print("No temperature parameter found")
# Use the model for inference
# ... your inference code here ...
```
## Training Details
- Base model: CLIP
- Custom temperature scaling included
- Fine-tuned for specific vision-language tasks
|