This is a fine-tuned version of Qwen/Qwen2.5-Coder-3B-Instruct for generating SQL queries from natural language questions. The model was fine-tuned using LoRA (r=16) on a subset of the Spider dataset and merged into a standalone model, eliminating the need for the peft library during inference. Usage To use the model for SQL query generation: from transformers import AutoModelForCausalLM, AutoTokenizer import torch

Load model and tokenizer

model_name = "Piyush026/Qwen2.5-Coder-3B-sql-finetuned" # Replace with your repo ID tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForCausalLM.from_pretrained( model_name, torch_dtype=torch.float16, device_map="auto", trust_remote_code=True )

Generate SQL query Example

prompt = """ Database: university Schema:

  • students: [student_id, first_name, last_name, department_code, gpa, major]
  • departments: [department_code, department_name]
  • courses: [course_number, course_title, professor_id]
  • instructors: [professor_id, last_name] Question: List all students. """ inputs = tokenizer(prompt, return_tensors="pt").to("cuda") outputs = model.generate(**inputs, max_length=200) print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Training Details

Base Model: Qwen/Qwen2.5-Coder-3B-Instruct Fine-Tuning: LoRA (r=16, lora_alpha=32, lora_dropout=0.05) on a 1000-sample subset of the Spider dataset. Environment: Lightning AI Studio with Tesla T4 GPU. Merged Model: The LoRA adapters were merged into the base model using merge_and_unload for standalone inference.

Downloads last month
10
Safetensors
Model size
3.09B params
Tensor type
F16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Piyush026/Qwen2.5-Coder-3B-sql-finetuned

Base model

Qwen/Qwen2.5-3B
Finetuned
(59)
this model