This is a fine-tuned version of Qwen/Qwen2.5-Coder-3B-Instruct for generating SQL queries from natural language questions. The model was fine-tuned using LoRA (r=16) on a subset of the Spider dataset and merged into a standalone model, eliminating the need for the peft library during inference. Usage To use the model for SQL query generation: from transformers import AutoModelForCausalLM, AutoTokenizer import torch
Load model and tokenizer
model_name = "Piyush026/Qwen2.5-Coder-3B-sql-finetuned" # Replace with your repo ID tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForCausalLM.from_pretrained( model_name, torch_dtype=torch.float16, device_map="auto", trust_remote_code=True )
Generate SQL query Example
prompt = """ Database: university Schema:
- students: [student_id, first_name, last_name, department_code, gpa, major]
- departments: [department_code, department_name]
- courses: [course_number, course_title, professor_id]
- instructors: [professor_id, last_name] Question: List all students. """ inputs = tokenizer(prompt, return_tensors="pt").to("cuda") outputs = model.generate(**inputs, max_length=200) print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Training Details
Base Model: Qwen/Qwen2.5-Coder-3B-Instruct Fine-Tuning: LoRA (r=16, lora_alpha=32, lora_dropout=0.05) on a 1000-sample subset of the Spider dataset. Environment: Lightning AI Studio with Tesla T4 GPU. Merged Model: The LoRA adapters were merged into the base model using merge_and_unload for standalone inference.
- Downloads last month
- 10