visual-qa-tem / README.md
LabSmart's picture
Update README.md
01c1341 verified
---
license: mit
metrics:
- bleu
- rouge
- meteor
- bertscore
base_model:
- liuhaotian/llava-v1.5-7b
pipeline_tag: visual-question-answering
---
# visual-qa-tem Model Card
## Model details
**base_model**
We finetune our custom data on LLava-v1.5-7b
See on :[liuhaotian/llava-v1.5-7b](https://huggingface.co/liuhaotian/llava-v1.5-7b)
**Paper or resources for more information:**
Our source code publish on : https://github.com/SmartLab-Roy/visual-qa-tem.git
### Download Model
```python
from huggingface_hub import snapshot_download
import os
# Download the model to local directory
model_path = snapshot_download(
repo_id="LabSmart/visual-qa-tem",
cache_dir="./models", # Local cache directory
resume_download=True
)
print(f"Model downloaded to: {model_path}")
```
### Quick Start
Reference [LLaVA](https://github.com/haotian-liu/LLaVA.git) for environment setup and CLI inference:
```
python -m llava.serve.cli \
--model-path "model_path from the download output"\
--image-file "path/to/your/tem_image.jpg" \
--load-4bit
```