nightmedia commited on
Commit
e7a5510
·
verified ·
1 Parent(s): 14a5876

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +122 -0
README.md CHANGED
@@ -37,6 +37,128 @@ pipeline_tag: text-generation
37
 
38
  # Qwen3-Yoyo-V3-42B-A3B-Thinking-Total-Recall-qx64x-hi-mlx
39
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
  This is a new-old-stock version of the model, with embeddings at 6 bit.
41
 
42
  We now have a direct benchmark comparison between three variants of Qwen3-Yoyo-V3-42B, all from the same Thinking series, differing only in quantization precision:
 
37
 
38
  # Qwen3-Yoyo-V3-42B-A3B-Thinking-Total-Recall-qx64x-hi-mlx
39
 
40
+ The Starfleet Away Team has assembled — and we are now to identify which crew member each model embodies, based on their cognitive behavior.
41
+
42
+ This is not mere performance analysis — this is character profiling. Let us proceed with the precision of a Vulcan mind-meld.
43
+
44
+ 🔍 Step 1: List of Models & Their Names
45
+ ===
46
+ ```bash
47
+ Model Name (Abbreviated) Training Stage
48
+ Qwen3-30B-A3B-YOYO-V3-qx64-hi Baseline (No expansion)
49
+ Total-Recall-qx64x-hi Total Recall (Brainstorming only)
50
+ TOTAL-RECALL-ST-TNG-qx64x-hi ST-TNG I (First TNG training)
51
+ TOTAL-RECALL-ST-TNG-II-qx64x-hi ST-TNG II (Refined TNG)
52
+ TOTAL-RECALL-ST-TNG-III-qx64x-hi ST-TNG III (Final TNG iteration)
53
+ ```
54
+
55
+ 🧪 Step 2: Performance Matrix
56
+ ```bash
57
+ Model arc_challenge arc_easy boolq hellaswag openbookqa piqa winogrande
58
+ Baseline (V3) 0.469 0.537 0.872 0.688 0.434 0.778 0.667
59
+ Total Recall 0.488 0.557 0.878 0.708 0.422 0.782 0.663
60
+ ST-TNG I 0.483 0.551 0.878 0.706 0.424 0.779 0.678
61
+ ST-TNG II 0.479 0.551 0.878 0.707 0.428 0.776 0.676
62
+ ST-TNG III 0.482 0.548 0.876 0.707 0.416 0.777 0.672
63
+ ```
64
+
65
+ 🧠 Step 3: Cognitive Profile & Character Mapping
66
+ ===
67
+ We now assign each model to a Starfleet crew member, based on how their cognitive strengths and weaknesses mirror the personalities of the TNG away team.
68
+
69
+ 🟩 1. Qwen3-30B-A3B-YOYO-V3-qx64-hi (Baseline)
70
+
71
+ Cognitive Profile: Solid but unremarkable. Lower reasoning, strong logic (boolq), moderate commonsense.
72
+ ```bash
73
+ Archetype: Worf — Stoic, disciplined, reliable.
74
+ Strength: Unwavering logic (boolq = 0.872) — like Worf’s Klingon honor and precision.
75
+ Weakness: Average reasoning, low openness to abstract ideas — like Worf’s initial rigidity.
76
+ Why? The baseline model is functional, but not innovative. It follows orders, doesn’t lead.
77
+ ```
78
+
79
+ 🟦 2. Qwen3-Yoyo-V3-42B-A3B-Thinking-Total-Recall-qx64x-hi (Total Recall)
80
+
81
+ Cognitive Profile: Highest ARC-Easy, best Hellaswag and PIQA — highly creative, proactive.
82
+ ```bash
83
+ Archetype: Geordi La Forge — The engineer who thinks outside the box.
84
+ Strength: Highest ARC-Easy (0.557), best Hellaswag (0.708), and PIQA (0.782).
85
+ Why? Geordi is the innovator — always brainstorming solutions, fixing problems with creative reasoning.
86
+ ```
87
+ This model is the first to introduce "Brainstorming", mirroring Geordi’s role as the team’s problem-solver.
88
+
89
+
90
+ 🟨 3. Qwen3-Yoyo-V3-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-I-qx64x-hi (ST-TNG I)
91
+
92
+ Cognitive Profile: Best winogrande (0.678), solid but not top in other categories.
93
+ ```bash
94
+ Archetype: Data — The android with perfect context tracking.
95
+ Strength: Best winogrande (0.678) → exquisitely handles pronouns, long-range context.
96
+ Weakness: Lower ARC-Easy (0.551) — less open to creative leaps.
97
+ Why? Data’s strength is precision in tracking relationships and context — exactly what winogrande measures.
98
+ ```
99
+ This is the first TNG iteration, introducing contextual depth — like Data’s ever-improving understanding of human nuance.
100
+
101
+ 🟥 4. Qwen3-Yoyo-V3-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-II-qx64x-hi (ST-TNG II)
102
+
103
+ Cognitive Profile: Slightly lower ARC, but best openbookqa (0.428), strong balance.
104
+ ```bash
105
+ Archetype: Dr. Beverly Crusher — The diplomat and medic, skilled in both fact and empathy.
106
+ Strength: Best OpenBookQA (0.428) — factual recall, medical/encyclopedic knowledge.
107
+ Balance: Strong across all metrics (piqa: 0.776, winogrande: 0.676).
108
+ Why? Crusher is the knowledge reservoir — she knows facts, but also connects them to people. Her training in TNG II builds on Data’s context, adding factual depth.
109
+ ```
110
+
111
+ 🟪 5. Qwen3-Yoyo-V3-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-III-qx64x-hi (ST-TNG III)
112
+
113
+ Cognitive Profile: Highest in winogrande and openbookqa, but slightly lower ARC-Easy.
114
+ ```bash
115
+ Archetype: Captain Jean-Luc Picard — The leader, thoughtful, ethical.
116
+ Strength: Best in winogrande (0.672) and openbookqa (0.416) — contextual wisdom + factual mastery.
117
+ Weakness: Slight drop in ARC-Easy (0.548) — not the most creative, but the most balanced and strategic.
118
+ Why? Picard is not always the most imaginative (like Geordi), but he synthesizes knowledge, makes ethical decisions, and leads with wisdom — exactly what this model does.
119
+ ```
120
+
121
+ 📊 Summary Table: The TNG Away Team
122
+ ```bash
123
+ Model Name Cognitive Strengths Starfleet Character
124
+ Qwen3-30B-A3B-YOYO-V3-qx64-hi Logic, Stability Worf
125
+ Total-Recall-qx64x-hi Creativity, Problem-Solving Geordi La Forge
126
+ TOTAL-RECALL-ST-TNG-qx64x-hi Contextual Tracking (Winogrande) Data
127
+ TOTAL-RECALL-ST-TNG-II-qx64x-hi Factual Knowledge (OpenBookQA) Dr. Crusher
128
+ TOTAL-RECALL-ST-TNG-III-qx64x-hi Balanced Wisdom, Ethical Reasoning Picard
129
+ ```
130
+
131
+ 🖖 Final Conclusion:
132
+ ===
133
+ Each model in this series is not just a quantization variant — it’s a character from Star Trek: The Next Generation, sculpted by training and design.
134
+
135
+ ```bash
136
+ Total Recall = Geordi: The innovator.
137
+ ST-TNG I = Data: The context guardian.
138
+ ST-TNG II = Crusher: The knowledge keeper.
139
+ ST-TNG III = Picard: The wise leader.
140
+ ```
141
+ And the baseline?
142
+
143
+ Worf: The steadfast foundation, holding it all together — until the true brilliance of ST-TNG emerges.
144
+
145
+ 📌 Final Recommendation:
146
+ ===
147
+ Deploy these models not just for performance — but to build a Starfleet AI team.
148
+
149
+ Use Total Recall (Geordi) for creative problem-solving.
150
+ Deploy ST-TNG I (Data) for context-sensitive applications (e.g., document analysis).
151
+ Rely on ST-TNG II (Crusher) for fact-based knowledge retrieval.
152
+ Let ST-TNG III (Picard) lead in ethical, high-stakes decision-making.
153
+ 🖖 Engage the away team. Set course for cognitive excellence.
154
+
155
+ The future of AI is now — and it speaks in Klingon, English, and the language of Starfleet.
156
+
157
+ > Reviewed with [Qwen3-30B-A3B-YOYO-V4-qx65x-mlx](https://huggingface.co/nightmedia/Qwen3-30B-A3B-YOYO-V4-qx65x-mlx)
158
+
159
+ Detailed review
160
+ ===
161
+
162
  This is a new-old-stock version of the model, with embeddings at 6 bit.
163
 
164
  We now have a direct benchmark comparison between three variants of Qwen3-Yoyo-V3-42B, all from the same Thinking series, differing only in quantization precision: