Italia-GPT
Italia-GPT is a 1.2B parameter model fine-tuned for native Italian instruction following. By focusing on linguistic nuances and cultural context, this model provides superior fluency compared to standard base models.
๐ Performance Overview
Below are the target benchmarks for the CALAMITA and Evalita-LLM frameworks:
| Metric | Score | Description |
|---|---|---|
| Logic & Reasoning | % | Native Italian sentence logic |
| Grammar Accuracy | % | Gender/Number agreement precision |
| Sentiment (ITA) | % | Detection of Italian irony and tone |
| Cultural Q&A | % | Localized knowledge and trivia |
๐ Technical Specifications
- Base Architecture: Llama 3.2
- Precision: BFloat16 ($BF16$)
- Weights: Merged Safetensors (Standalone)
- Language Support: Primary: Italian ๐ฎ๐น, Secondary: English ๐บ๐ธ
๐ Usage Guide
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
model_id = "XythicK/Italia-GPT"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
model_id,
torch_dtype=torch.bfloat16,
device_map="auto"
)
# Native Italian Chat Template
messages = [{"role": "user", "content": "Come si prepara una vera carbonara?"}]
inputs = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt").to("cuda")
outputs = model.generate(inputs, max_new_tokens=256)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
- Downloads last month
- 22