Canis.teach โ€” Qwen3โ€‘4B Instruct (Generalist) โ€” Merged

Merged full model (LoRA adapters applied to the base), ready for direct use with Transformers.

  • Base: Qwen/Qwen3-4B-Instruct-2507
  • Release: CanisAI/teach-generalist-qwen3-4b-2507-r1-merged
  • Project: Canis.teach, Learning that fits.
  • Tags: canis-teach, qwen3, education, lora-merged, transformers

What is this?

This repository contains a merged checkpoint: the LoRA adapters fineโ€‘tuned on Generalist tutoring dialogues have been merged into the base model (Qwen/Qwen3โ€‘4Bโ€‘Instructโ€‘2507). This allows you to load and run the model directly with Transformers (no PEFT merge step at runtime).

For lightweight adapters or Ollama-friendly quantized builds, see the โ€œRelatedโ€ section.

Quick usage (Transformers)

from transformers import AutoTokenizer, AutoModelForCausalLM

repo = "CanisAI/teach-generalist-qwen3-4b-2507-r1-merged"

tok = AutoTokenizer.from_pretrained(repo, use_fast=True)
model = AutoModelForCausalLM.from_pretrained(
    repo,
    device_map="auto",
    torch_dtype="auto"
)

prompt = "Explain how to solve 2x + 1 = 5 step by step."
inputs = tok(prompt, return_tensors="pt").to(model.device)
out = model.generate(**inputs, max_new_tokens=256, temperature=0.7, top_p=0.8, top_k=20)
print(tok.decode(out[0], skip_special_tokens=True))

Recommended decoding (for instruct-style usage):

  • temperature โ‰ˆ 0.7
  • top_p โ‰ˆ 0.8
  • top_k โ‰ˆ 20 Adjust to your needs.

Intended use

  • Subjectโ€‘aware tutoring for Generalist with didactic, stepโ€‘byโ€‘step responses.
  • Suitable for educational prototypes, demonstrations, and research.
  • Built to โ€œteach, not just answerโ€: stepwise hints, clarity, and rubricโ€‘aligned structure.

Safety and limitations

  • Human oversight is required. The model may hallucinate or oversimplify.
  • For factโ€‘heavy tasks, consider Retrievalโ€‘Augmented Generation (RAG) with curriculum sources.
  • Follow data privacy and compliance rules in your environment (e.g., school policies).

Training summary

  • Base model: Qwen/Qwen3-4B-Instruct-2507
  • Method: Supervised fineโ€‘tuning with LoRA (Unsloth + TRL/PEFT), then merged to full weights
  • Data: Subjectโ€‘specific tutoring dialogues generated/curated via Canis.lab
  • Goal: Improve clarity, hints, and step-by-step pedagogy for Generalist

Note: Exact hyperparameters and logs are provided in the LoRA training pipeline (if published) or available on request.

Related

  • LoRA adapters (lightweight):
    • CanisAI/teach-generalist-qwen3-4b-2507-r1
  • Quantized GGUF for Ollama/llama.cpp:
    • CanisAI/teach-generalist-qwen3-4b-2507-r1-gguf
  • Base model:
    • Qwen/Qwen3-4B-Instruct-2507

License

  • Inherits the base modelโ€™s license. Review the base model terms before use.
  • Dataset licensing and any thirdโ€‘party assets should be respected accordingly.

Acknowledgments

  • Qwen3 by Qwen team
  • Unsloth, TRL, PEFT, and Transformers for training/serving
  • Educators and contributors supporting Canis.teach

Learning that fits.

Downloads last month
31
Safetensors
Model size
4B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for CanisAI/teach-generalist-qwen3-4b-2507-r1-merged

Merge model
this model

Dataset used to train CanisAI/teach-generalist-qwen3-4b-2507-r1-merged

Space using CanisAI/teach-generalist-qwen3-4b-2507-r1-merged 1