File size: 1,076 Bytes
da78f41
b201b46
 
 
 
 
 
da78f41
 
b201b46
da78f41
b201b46
da78f41
b201b46
da78f41
b201b46
da78f41
b201b46
da78f41
b201b46
 
 
da78f41
b201b46
 
 
da78f41
b201b46
 
 
 
 
 
 
 
da78f41
b201b46
 
 
da78f41
 
 
b201b46
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
---
tags:
- clip
- vision-language
- image-text
- pytorch
license: apache-2.0
---

# CLIP Model

This is a fine-tuned CLIP model for vision-language tasks.

## Model Description

This model was fine-tuned from a base CLIP model and includes custom temperature scaling.

## Usage

```python
from transformers import CLIPModel, CLIPProcessor
import torch

# Load model and processor
model = CLIPModel.from_pretrained("aprendesc/CLIP_model_v0")
processor = CLIPProcessor.from_pretrained("aprendesc/CLIP_model_v0")

# Load temperature parameter if available
try:
    from huggingface_hub import hf_hub_download
    temperature_path = hf_hub_download(repo_id="aprendesc/CLIP_model_v0", filename="temperature.pth")
    temperature = torch.load(temperature_path, map_location='cpu')
    print(f"Temperature parameter: {temperature}")
except:
    print("No temperature parameter found")

# Use the model for inference
# ... your inference code here ...
```

## Training Details

- Base model: CLIP
- Custom temperature scaling included
- Fine-tuned for specific vision-language tasks