Italia-GPT

Italia-GPT is a 1.2B parameter model fine-tuned for native Italian instruction following. By focusing on linguistic nuances and cultural context, this model provides superior fluency compared to standard base models.

Model Card Model Size


๐Ÿ’Ž Performance Overview

Below are the target benchmarks for the CALAMITA and Evalita-LLM frameworks:

Metric Score Description
Logic & Reasoning % Native Italian sentence logic
Grammar Accuracy % Gender/Number agreement precision
Sentiment (ITA) % Detection of Italian irony and tone
Cultural Q&A % Localized knowledge and trivia

๐Ÿ›  Technical Specifications

  • Base Architecture: Llama 3.2
  • Precision: BFloat16 ($BF16$)
  • Weights: Merged Safetensors (Standalone)
  • Language Support: Primary: Italian ๐Ÿ‡ฎ๐Ÿ‡น, Secondary: English ๐Ÿ‡บ๐Ÿ‡ธ

๐Ÿš€ Usage Guide

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

model_id = "XythicK/Italia-GPT"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
    model_id, 
    torch_dtype=torch.bfloat16, 
    device_map="auto"
)

# Native Italian Chat Template
messages = [{"role": "user", "content": "Come si prepara una vera carbonara?"}]
inputs = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt").to("cuda")
outputs = model.generate(inputs, max_new_tokens=256)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Downloads last month
22
Safetensors
Model size
1B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for XythicK/Italia-GPT

Finetuned
(1423)
this model
Quantizations
2 models