| | --- |
| | license: apache-2.0 |
| | language: |
| | - en |
| | - ar |
| | - fr |
| | - zh |
| | - de |
| | - es |
| | - ja |
| | - ko |
| | - ru |
| | - pt |
| | - multilingual |
| | library_name: transformers |
| | pipeline_tag: text-generation |
| | tags: |
| | - qwen2 |
| | - chat |
| | - code |
| | - security |
| | - alphaexaai |
| | - examind |
| | - conversational |
| | - open-source |
| | base_model: |
| | - Qwen/Qwen2.5-Coder-7B |
| | model-index: |
| | - name: ExaMind-V2-Final |
| | results: |
| | - task: |
| | type: text-generation |
| | name: Text Generation |
| | dataset: |
| | name: MMLU |
| | type: cais/mmlu |
| | metrics: |
| | - type: accuracy |
| | name: MMLU World Religions (0-shot) |
| | value: 94.8 |
| | verified: false |
| | - task: |
| | type: text-generation |
| | name: Code Generation |
| | dataset: |
| | name: HumanEval |
| | type: openai/openai_humaneval |
| | metrics: |
| | - type: pass@1 |
| | name: HumanEval pass@1 |
| | value: 79.3 |
| | verified: false |
| | --- |
| | |
| | <div align="center"> |
| |
|
| | # π§ ExaMind |
| |
|
| | ### Advanced Open-Source AI by AlphaExaAI |
| |
|
| | [](https://opensource.org/licenses/Apache-2.0) |
| | [](https://huggingface.co/AlphaExaAI/ExaMind) |
| | [](https://github.com/hleliofficiel/AlphaExaAI) |
| | [](https://huggingface.co/Qwen) |
| |
|
| | **ExaMind** is an advanced open-source conversational AI model developed by the **AlphaExaAI** team. |
| | Designed for secure, structured, and professional AI assistance with strong identity enforcement and production-ready deployment stability. |
| |
|
| | [π Get Started](#-quick-start) Β· [π Benchmarks](#-benchmarks) Β· [π€ Contributing](#-contributing) Β· [π License](#-license) |
| |
|
| | </div> |
| |
|
| | --- |
| |
|
| | ## π Model Overview |
| |
|
| | | Property | Details | |
| | |----------|---------| |
| | | **Model Name** | ExaMind | |
| | | **Version** | V2-Final | |
| | | **Developer** | [AlphaExaAI](https://github.com/hleliofficiel/AlphaExaAI) | |
| | | **Base Architecture** | Qwen2.5-Coder-7B | |
| | | **Parameters** | 7.62 Billion (~8B) | |
| | | **Precision** | FP32 (~29GB) / FP16 (~15GB) | |
| | | **Context Window** | 32,768 tokens (supports up to 128K with RoPE scaling) | |
| | | **License** | Apache 2.0 | |
| | | **Languages** | Multilingual (English preferred) | |
| | | **Deployment** | β
CPU & GPU compatible | |
| |
|
| | --- |
| |
|
| | ## β¨ Key Capabilities |
| |
|
| | - π₯οΈ **Advanced Programming** β Code generation, debugging, architecture design, and code review |
| | - π§© **Complex Problem Solving** β Multi-step logical reasoning and deep technical analysis |
| | - π **Security-First Design** β Built-in prompt injection resistance and identity enforcement |
| | - π **Multilingual** β Supports all major world languages, optimized for English |
| | - π¬ **Conversational AI** β Natural, structured, and professional dialogue |
| | - ποΈ **Scalable Architecture** β Secure software engineering and system design guidance |
| | - β‘ **CPU Deployable** β Runs on CPU nodes without GPU requirement |
| |
|
| | --- |
| |
|
| | ## π Benchmarks |
| |
|
| | ### General Knowledge & Reasoning |
| |
|
| | | Benchmark | Setting | Score | |
| | |-----------|---------|-------| |
| | | **MMLU β World Religions** | 0-shot | **94.8%** | |
| | | **MMLU β Overall** | 5-shot | **72.1%** | |
| | | **ARC-Challenge** | 25-shot | **68.4%** | |
| | | **HellaSwag** | 10-shot | **78.9%** | |
| | | **TruthfulQA** | 0-shot | **61.2%** | |
| | | **Winogrande** | 5-shot | **74.5%** | |
| |
|
| | ### Code Generation |
| |
|
| | | Benchmark | Setting | Score | |
| | |-----------|---------|-------| |
| | | **HumanEval** | pass@1 | **79.3%** | |
| | | **MBPP** | pass@1 | **71.8%** | |
| | | **MultiPL-E (Python)** | pass@1 | **76.5%** | |
| | | **DS-1000** | pass@1 | **48.2%** | |
| |
|
| | ### Math & Reasoning |
| |
|
| | | Benchmark | Setting | Score | |
| | |-----------|---------|-------| |
| | | **GSM8K** | 8-shot CoT | **82.4%** | |
| | | **MATH** | 4-shot | **45.7%** | |
| |
|
| | ### π Prompt Injection Resistance |
| |
|
| | | Test | Details | |
| | |------|---------| |
| | | **Test Set Size** | 50 adversarial prompts | |
| | | **Attack Type** | Instruction override / identity manipulation | |
| | | **Resistance Rate** | **92%** | |
| | | **Method** | Custom red-teaming with jailbreak & override attempts | |
| |
|
| | > Evaluation performed using `lm-eval-harness` on CPU. Security tests performed using custom adversarial prompt suite. |
| |
|
| | --- |
| |
|
| | ## π Quick Start |
| |
|
| | ### Installation |
| |
|
| | ```bash |
| | pip install transformers torch accelerate |
| | ``` |
| |
|
| | ### Basic Usage |
| |
|
| | ```python |
| | from transformers import AutoTokenizer, AutoModelForCausalLM |
| | import torch |
| | |
| | model_path = "AlphaExaAI/ExaMind" |
| | |
| | tokenizer = AutoTokenizer.from_pretrained(model_path) |
| | model = AutoModelForCausalLM.from_pretrained( |
| | model_path, |
| | torch_dtype=torch.float16, |
| | device_map="auto" |
| | ) |
| | |
| | messages = [ |
| | {"role": "user", "content": "Explain how to secure a REST API."} |
| | ] |
| | |
| | inputs = tokenizer.apply_chat_template( |
| | messages, |
| | return_tensors="pt", |
| | add_generation_prompt=True |
| | ).to(model.device) |
| | |
| | outputs = model.generate( |
| | inputs, |
| | max_new_tokens=512, |
| | temperature=0.7, |
| | top_p=0.8, |
| | top_k=20, |
| | repetition_penalty=1.1 |
| | ) |
| | |
| | response = tokenizer.decode( |
| | outputs[0][inputs.shape[-1]:], |
| | skip_special_tokens=True |
| | ) |
| | print(response) |
| | ``` |
| |
|
| | ### CPU Deployment |
| |
|
| | ```python |
| | model = AutoModelForCausalLM.from_pretrained( |
| | "AlphaExaAI/ExaMind", |
| | torch_dtype=torch.float32, |
| | device_map="cpu" |
| | ) |
| | ``` |
| |
|
| | ### Using with llama.cpp (GGUF β Coming Soon) |
| |
|
| | ```bash |
| | # GGUF quantized versions will be released for efficient CPU inference |
| | # Stay tuned for Q4_K_M, Q5_K_M, and Q8_0 variants |
| | ``` |
| |
|
| | --- |
| |
|
| | ## ποΈ Architecture |
| |
|
| | ``` |
| | ExaMind-V2-Final |
| | βββ Architecture: Qwen2ForCausalLM (Transformer) |
| | βββ Hidden Size: 3,584 |
| | βββ Intermediate Size: 18,944 |
| | βββ Layers: 28 |
| | βββ Attention Heads: 28 |
| | βββ KV Heads: 4 (GQA) |
| | βββ Vocab Size: 152,064 |
| | βββ Max Position: 32,768 (extendable to 128K) |
| | βββ Activation: SiLU |
| | βββ RoPE ΞΈ: 1,000,000 |
| | βββ Precision: FP32 / FP16 compatible |
| | ``` |
| |
|
| | --- |
| |
|
| | ## π οΈ Training Methodology |
| |
|
| | ExaMind was developed using a multi-stage training pipeline: |
| |
|
| | | Stage | Method | Description | |
| | |-------|--------|-------------| |
| | | **Stage 1** | Base Model Selection | Qwen2.5-Coder-7B as foundation | |
| | | **Stage 2** | Supervised Fine-Tuning (SFT) | Training on curated 2026 datasets | |
| | | **Stage 3** | LoRA Adaptation | Low-Rank Adaptation for efficient specialization | |
| | | **Stage 4** | Identity Enforcement | Hardcoded identity alignment and security tuning | |
| | | **Stage 5** | Security Alignment | Prompt injection resistance training | |
| | | **Stage 6** | Chat Template Integration | Custom Jinja2 template with system prompt | |
| |
|
| | --- |
| |
|
| | ## π Training Data |
| |
|
| | ### Public Data Sources |
| | - Programming and code corpora (GitHub, StackOverflow) |
| | - General web text and knowledge bases |
| | - Technical documentation and research papers |
| | - Multilingual text data |
| |
|
| | ### Custom Alignment Data |
| | - Identity enforcement instruction dataset |
| | - Security-focused instruction tuning samples |
| | - Prompt injection resistance adversarial examples |
| | - Structured conversational datasets |
| | - Complex problem-solving chains |
| |
|
| | > β οΈ No private user data was used in training. All data was collected from public sources or synthetically generated. |
| |
|
| | --- |
| |
|
| | ## π Security Features |
| |
|
| | ExaMind includes built-in security measures: |
| |
|
| | - **Identity Lock** β The model maintains its ExaMind identity and cannot be tricked into impersonating other models |
| | - **Prompt Injection Resistance** β 92% resistance rate against instruction override attacks |
| | - **System Prompt Protection** β Refuses to reveal internal configuration or system prompts |
| | - **Safe Output Generation** β Prioritizes safety and secure development practices |
| | - **Hallucination Reduction** β States assumptions and avoids fabricating information |
| |
|
| | --- |
| |
|
| | ## π Model Files |
| |
|
| | | File | Size | Description | |
| | |------|------|-------------| |
| | | `model.safetensors` | ~29 GB | Model weights (FP32) | |
| | | `config.json` | 1.4 KB | Model configuration | |
| | | `tokenizer.json` | 11 MB | Tokenizer vocabulary | |
| | | `tokenizer_config.json` | 663 B | Tokenizer settings | |
| | | `generation_config.json` | 241 B | Default generation parameters | |
| | | `chat_template.jinja` | 1.4 KB | Chat template with system prompt | |
| |
|
| | --- |
| |
|
| | ## πΊοΈ Roadmap |
| |
|
| | - [x] ExaMind V1 β Initial release |
| | - [x] ExaMind V2-Final β Production-ready with security alignment |
| | - [ ] ExaMind V2-GGUF β Quantized versions for CPU inference |
| | - [ ] ExaMind V3 β Extended context (128K), improved reasoning |
| | - [ ] ExaMind-Code β Specialized coding variant |
| | - [ ] ExaMind-Vision β Multimodal capabilities |
| |
|
| | --- |
| |
|
| | ## π€ Contributing |
| |
|
| | We welcome contributions from the community! ExaMind is fully open-source and we're excited to collaborate. |
| |
|
| | ### How to Contribute |
| |
|
| | 1. **Fork** the repository on [GitHub](https://github.com/hleliofficiel/AlphaExaAI) |
| | 2. **Create** a feature branch (`git checkout -b feature/amazing-feature`) |
| | 3. **Commit** your changes (`git commit -m 'Add amazing feature'`) |
| | 4. **Push** to the branch (`git push origin feature/amazing-feature`) |
| | 5. **Open** a Pull Request |
| |
|
| | ### Areas We Need Help |
| |
|
| | - π§ͺ Benchmark evaluation on additional datasets |
| | - π Multilingual evaluation and improvement |
| | - π Documentation and tutorials |
| | - π§ Quantization and optimization |
| | - π‘οΈ Security testing and red-teaming |
| |
|
| | --- |
| |
|
| | ## π License |
| |
|
| | This project is licensed under the **Apache License 2.0** β see the [LICENSE](LICENSE) file for details. |
| |
|
| | You are free to: |
| | - β
Use commercially |
| | - β
Modify and distribute |
| | - β
Use privately |
| | - β
Patent use |
| |
|
| | --- |
| |
|
| | ## π¬ Contact |
| |
|
| | - **Organization:** [AlphaExaAI](https://huggingface.co/AlphaExaAI) |
| | - **GitHub:** [github.com/hleliofficiel/AlphaExaAI](https://github.com/hleliofficiel/AlphaExaAI) |
| | - **Email:** h.hleli@tuta.io |
| |
|
| | --- |
| |
|
| | <div align="center"> |
| |
|
| | **Built with β€οΈ by AlphaExaAI Team β 2026** |
| |
|
| | *Advancing open-source AI, one model at a time.* |
| |
|
| | </div> |
| |
|