lainanon-0.6b

A fine tune of Qwen/Qwen3-0.6B, a 600M parameter causal language model. Fine-tuned on the rudyon/lainchan-alpaca dataset, a scraped collection of posts from lainchan.org, using LoRA fine-tuning with bfloat16 precision on a Vast.ai instance with a single RTX 5090.

usage

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
import torch

base = AutoModelForCausalLM.from_pretrained("Qwen/Qwen3-0.6B", torch_dtype=torch.bfloat16, device_map="auto")
tokenizer = AutoTokenizer.from_pretrained("rudyon/lainanon-0.6b")
model = PeftModel.from_pretrained(base, "rudyon/lainanon-0.6b")

def chat(instruction, max_new_tokens=200):
    prompt = f"### Instruction:\n{instruction}\n### Response:\n"
    inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
    outputs = model.generate(**inputs, max_new_tokens=max_new_tokens)
    return tokenizer.decode(outputs[0], skip_special_tokens=True)

print(chat("How do I get started with Emacs?"))

prompt format

### Instruction:
Your instruction here
### Response:
Downloads last month
51
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for rudyon/lainanon-0.6b

Finetuned
Qwen/Qwen3-0.6B
Adapter
(368)
this model

Dataset used to train rudyon/lainanon-0.6b