Configuration Parsing Warning: In adapter_config.json: "peft.task_type" must be a string

GPT-OSS-20B Empathetic (LoRA Fine-tuned)

This model is a LoRA fine-tuned adapter built on top of unsloth/gpt-oss-20b-unsloth-bnb-4bit.
It specializes in generating empathetic and supportive responses, making it suitable for conversational AI use cases where emotional awareness is important.

Model Details

Model Description

  • Developed by: Anwesha026
  • Shared by: Anwesha026
  • Model type: Decoder-only Causal LM (LoRA adapter)
  • Language(s): English
  • License: Apache-2.0
  • Finetuned from model [optional]: unsloth/gpt-oss-20b-unsloth-bnb-4bit

Model Sources

Uses

Direct Use

  • Empathetic chatbots
  • Companion-like conversational assistants
  • Research in affective computing and emotionally aware dialogue

Downstream Use

  • Integration into mental health support tools (with human supervision)
  • Conversational agents requiring emotionally supportive responses

Out-of-Scope Use

  • Providing professional medical or psychological advice
  • Factual Q&A where high accuracy is required
  • Malicious or manipulative applications

Bias, Risks, and Limitations

Like most LLMs, this model may:

  • Produce biased, stereotypical, or culturally insensitive outputs
  • Over-generalize empathetic responses
  • Hallucinate factual details
  • Fail in high-stakes or sensitive psychological contexts

Recommendations

  • Always keep a human in the loop when deploying in sensitive domains
  • Do not use as a replacement for professional medical/psychological help
  • Carefully evaluate outputs before real-world use

How to Get Started with the Model

Use the code below to get started with the model.

from transformers import AutoTokenizer, AutoModelForCausalLM
import torch
model_id = "Anwesha026/fine-tuned-gpt-oss-20b"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.bfloat16)

inputs = tokenizer("I feel really lonely lately.", return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=100)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Training Details

Training Data

Dataset: facebook/empathetic_dialogues

Training Procedure

Training Hyperparameters

  • Batch size (per device): 1

  • Gradient accumulation steps: 4 → effective batch size = 1 × 4 = 4

  • Learning rate: 1e-4

  • Optimizer: AdamW (8-bit)

  • Weight decay: 0.01

  • Learning rate scheduler: Linear

  • Warmup steps: 10

  • Max training steps: 300

  • Seed: 3407

Evaluation

Results

  • Improved empathetic alignment compared to the base model

  • Some generic/repetitive answers persist

Technical Specifications

Model Architecture and Objective

  • Base model: GPT-OSS-20B (decoder-only transformer, 20B parameters)

  • Fine-tuning method: LoRA adapters via PEFT

Compute Infrastructure

Hardware

  • NVIDIA GPU

Software

  • Hugging Face Transformers, PEFT, TRL, Unsloth

Model Card Authors

  • Anwesha026

Model Card Contact

Framework versions

  • Transformers: 4.x

  • PEFT: 0.17.1

  • TRL: latest

  • Unsloth: latest

Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Anwesha026/fine-tuned-gpt-oss-20b

Base model

openai/gpt-oss-20b
Adapter
(38)
this model