π IoraX 3B β Efficient Conversational AI Model
β¨ Model Overview
IoraX 3B is a highly efficient 3-billion parameter Transformer, fine-tuned using LoRA adapters on Meta LLaMA 3.2 (3B) β with 4-bit quantization to keep it lightning fast and lightweight!
This model specializes in deep conversational understanding, logical reasoning, and coherent long-form generation β your AI companion for research, education, and creative tasks.
π― Features & Capabilities
- π§ Size: 3B parameters
- βοΈ Base: Meta LLaMA 3.2 (3B)
- π§ Fine-tuning: LoRA with 4-bit quantization
- β³ Max context length: 2048 tokens (with RoPE scaling)
- π Training data: Blend of public conversational datasets + expert-curated Q&A
- π Epochs: 3 for balanced speed and learning
- π Language: English
π Use Cases
| Use Case | Description |
|---|---|
| π¬ Conversational AI | Customer support, chatbots, assistants |
| π Education | Tutoring, concept explanation, Q&A |
| π§ͺ Research Assistant | Drafting, summarizing, brainstorming |
| βοΈ Creative Writing | Storytelling, script generation |
β οΈ Limitations
- π Knowledge cutoff: Data up to 2023 only
- βοΈ Bias: May reflect biases present in the training corpus
- βοΈ Accuracy: Verify important outputs, especially in critical domains
- π§ββοΈ Not a replacement for experts: Use responsibly
π‘ Quick Start
from transformers import AutoTokenizer
from unsloth import FastLanguageModel
model_name = "XythicK/IoraX-3B"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = FastLanguageModel.from_pretrained(model_name, load_in_4bit=True, max_seq_length=2048)
messages = [
{"role": "user", "content": "Explain the philosophical significance of the Eiffel Tower. ππ€"}
]
inputs = tokenizer.apply_chat_template(
messages,
tokenize=True,
add_generation_prompt=True,
return_tensors="pt"
).to("cuda")
outputs = model.generate(
input_ids=inputs,
max_new_tokens=128,
temperature=1.2,
use_cache=True
)
print(tokenizer.batch_decode(outputs, skip_special_tokens=True))
π Contact
Maintainer: M Mashhudur Rahim [XythicK]
Role:
Independent Machine Learning Researcher & Model Infrastructure Maintainer
(Focused on model quantization, optimization, and efficient deployment)
For issues, improvement requests, or additional quantization formats, please use the Hugging Face Discussions or Issues tab.
π Citation
If you use IoraX in your work, please cite:
@misc{ioraX2025,
title = {IoraX 3B: Efficient Conversational AI},
author = {M Mashhudur Rahim (XythicK)},
year = {2025},
publisher = {Hugging Face},
howpublished = {\url{https://huggingface.co/XythicK/IoraX-3B}}
}
β€οΈ Acknowledgements
Thanks to Hugging Face and the open-source machine learning community for providing the tools and platforms that make efficient model sharing and deployment possible.
- Downloads last month
- -
