prithivMLmods commited on
Commit
9f4ffbf
·
verified ·
1 Parent(s): 7166ec3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +111 -3
README.md CHANGED
@@ -1,3 +1,111 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model:
4
+ - prithivMLmods/Sombrero-Opus-14B-Elite6
5
+ pipeline_tag: text-generation
6
+ library_name: transformers
7
+ tags:
8
+ - text-generation-inference
9
+ - code
10
+ - math
11
+ - R1
12
+ language:
13
+ - en
14
+ ---
15
+
16
+ ![2.png](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/IUa2oAhsMWrD9qwTrEEAD.png)
17
+
18
+ # **La-Superba-14B-Y.2**
19
+
20
+ > **La-Superba-14B-Y.2** is a next-generation language model built on the Qwen 2.5 14B architecture. It is meticulously optimized for **mathematical reasoning**, **programming**, and **general-purpose logic-based tasks**. With its advanced comprehension, structured problem-solving capabilities, and long-context handling, it serves as a powerful assistant for technical, educational, and reasoning-intensive workflows.
21
+
22
+ ## **Key Improvements**
23
+ 1. **Exceptional Mathematical Reasoning**
24
+ Specially trained for handling symbolic math, arithmetic, algebra, calculus, and applied mathematics with step-by-step clarity and logical precision.
25
+
26
+ 2. **Advanced Coding & Debugging Intelligence**
27
+ Proficient in code generation, multi-language programming support (Python, JavaScript, C++, etc.), and automatic debugging. It can explain, optimize, and refactor code with minimal prompting.
28
+
29
+ 3. **Superior General-Purpose Reasoning**
30
+ Fine-tuned to manage logical deduction, multi-step reasoning, and contextual understanding across a wide array of domains.
31
+
32
+ 4. **Instruction-Following Accuracy**
33
+ Capable of precisely interpreting nested, multi-part instructions and returning structured, coherent responses that follow the prompt intent faithfully.
34
+
35
+ 5. **Extended Context Support**
36
+ Handles up to **128K tokens** of input with **8K token** output capacity, making it suitable for long documents, codebases, and detailed walkthroughs.
37
+
38
+
39
+ ## **Quickstart with Transformers**
40
+
41
+ ```python
42
+ from transformers import AutoModelForCausalLM, AutoTokenizer
43
+
44
+ model_name = "prithivMLmods/La-Superba-14B-Y.2"
45
+
46
+ model = AutoModelForCausalLM.from_pretrained(
47
+ model_name,
48
+ torch_dtype="auto",
49
+ device_map="auto"
50
+ )
51
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
52
+
53
+ prompt = "Write a Python function to check whether a number is prime and explain each step."
54
+ messages = [
55
+ {"role": "system", "content": "You are a highly capable assistant in math, programming, and logical reasoning."},
56
+ {"role": "user", "content": prompt}
57
+ ]
58
+ text = tokenizer.apply_chat_template(
59
+ messages,
60
+ tokenize=False,
61
+ add_generation_prompt=True
62
+ )
63
+ model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
64
+
65
+ generated_ids = model.generate(
66
+ **model_inputs,
67
+ max_new_tokens=512
68
+ )
69
+ generated_ids = [
70
+ output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
71
+ ]
72
+
73
+ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
74
+ ```
75
+
76
+ ## **Intended Use**
77
+
78
+ 1. **Mathematical Problem Solving**
79
+ Solves math questions with detailed steps, symbolic manipulation, and numerical precision—ideal for students, educators, and professionals.
80
+
81
+ 2. **Programming & Automation**
82
+ Assists in writing clean, correct code and helps debug and explain errors in software development tasks.
83
+
84
+ 3. **Technical Support and Tutoring**
85
+ Can be deployed as a tutor or assistant in educational platforms focused on logic, STEM, and engineering disciplines.
86
+
87
+ 4. **General-Purpose Reasoning Agent**
88
+ Useful in applications requiring thoughtful multi-turn reasoning, structured outputs, and logical consistency.
89
+
90
+ 5. **Multilingual Knowledge Assistant**
91
+ Enables intelligent communication and content generation across various languages and technical contexts.
92
+
93
+ 6. **Structured and Long-Form Output**
94
+ Can produce well-formatted JSON, tables, documents, and full-length guides and reports while maintaining coherence.
95
+
96
+ ## **Limitations**
97
+
98
+ 1. **High Hardware Demand**
99
+ Best performance requires high-RAM GPUs or TPUs due to its parameter size and context window.
100
+
101
+ 2. **Bias and Factual Limits**
102
+ Some inherited training data biases and occasional factual inaccuracies may still appear.
103
+
104
+ 3. **Not Real-Time Aware**
105
+ It does not have access to current events or real-time information post-training.
106
+
107
+ 4. **Creative Limitations**
108
+ Less consistent with storytelling, poetry, or heavily subjective tasks.
109
+
110
+ 5. **Prompt Sensitivity**
111
+ Output quality and structure can vary based on prompt clarity and format.