SmolLM-360M-CustomerSupport-Instruct

This is a fine-tuned version of HuggingFaceTB/SmolLM-360M-Instruct trained on customer support conversations.

Model Details

  • Base Model: SmolLM-360M-Instruct (360M parameters)
  • Fine-tuning Method: LoRA (Low-Rank Adaptation)
  • Dataset: Bitext Customer Support Dataset
  • Training Samples: 2,000 customer support conversations
  • Use Case: Customer service chatbot, support ticket handling

Training Details

  • Epochs: 3
  • Learning Rate: 1e-4
  • Batch Size: 8
  • Max Sequence Length: 512
  • Optimizer: AdamW with cosine learning rate schedule
  • Hardware: Single GPU (Colab/Kaggle)

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "sweatSmile/SmolLM-360M-CustomerSupport-Instruct"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

# Format your prompt
prompt = "<|im_start|>user\nHow do I reset my password?<|im_end|>\n<|im_start|>assistant\n"

inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=150, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)

print(response)

Example Outputs

Input: "My order hasn't arrived yet, what should I do?"

Output: "I apologize for the delay with your order. Let me help you track it. Could you please provide your order number? Once I have that, I can check the current status and estimated delivery date."

Limitations

  • Trained on English customer support conversations only
  • Best for customer service domains (banking, e-commerce, general support)
  • May not perform well on highly technical or domain-specific queries outside training data
  • Small model size (360M) - not as capable as larger models but much faster

Intended Use

  • Customer support chatbots
  • Automated ticket response systems
  • FAQ assistance
  • Initial customer inquiry handling

Training Data

The model was fine-tuned on the Bitext Customer Support Dataset, which contains diverse customer service scenarios.

Citation

@misc{smollm-customer-support-2025,
  author = {sweatSmile},
  title = {SmolLM-360M Fine-tuned for Customer Support},
  year = {2025},
  publisher = {HuggingFace},
  url = {https://huggingface.co/sweatSmile/SmolLM-360M-CustomerSupport-Instruct}
}

License

Apache 2.0 (same as base model)

Downloads last month
2
Safetensors
Model size
0.4B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for sweatSmile/SmolLM-360M-CustomerSupport-Instruct

Adapter
(8)
this model