Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up
reaperdoesntknow 
posted an update 4 days ago
Post
3228
We present a methodology for training small language models on CPU at FP32 precision
that achieves capability-per-dollar efficiency orders of magnitude beyond GPU-based training.
Across15modelsspanningfournovelarchitecturefamilies—MixtureofAttentions(MoA),cross-
architecture fusion (Qemma), swarm intelligence (SAGI), and metric-space causal language
models (DiscoverLM)—total compute cost was $24 on a single AMD EPYC 9454P proces-
sor. We introduce seven methodological pillars: (1) FP32 precision preservation, with exper-
iments demonstrating 5,810×single-operation error and 23,225×compounding error ratio for
FP16 at network depth; (2) sparse cognitive architectures where 0.02–7% of parameters activate
per token, matching CPU branching rather than GPU SIMD; (3) developmental curriculum
training progressing from language to logic to transfer to depth; (4) continuous belt-fed data
ingestion eliminating truncation waste; (5) hardware-native optimization for AMD Zen 4 via
AOCL/OpenMP/NUMA-aware allocation; (6) self-regulating thermodynamic governance with
emergent temperature measurement grounded in L2-star discrepancy; and (7) open-standard
compute (AVX2 SIMD at FP32) free of proprietary vendor dependency. We argue that transformers were designed for GPU hardware rather than mathematical optimality, and that architecture designed for geometric correctness—metric-space attention, triangle inequality enforcement, sparse expert routing—naturally favor CPU execution. For sub-2B parameter models, CPU training produces more capable models at a fraction of the cost.

@reaperdoesntknow Are the trained models and the training code opensourced?

·

All models are open-weight under Apache 2.0 — grab them directly from the repos. Training methodology is documented in our Structure Over Scale paper (DOI: 10.57967/hf/8165). Full training code release is coming alongside the next paper drop. Stay tuned.
https://huggingface.co/reaperdoesntknow/Structure-Over-Scale

Interesting, so why don't you create a research paper? Wanna see the training recipe and configuration, setup

·

Paper exists — Structure Over Scale (DOI: 10.57967/hf/8165) on HuggingFace. Covers the seven pillars, FP32 precision rationale, and the full 15-model empirical catalog. Training configs and hyperparameters are in the individual model cards. A deeper methodology paper (topology-aware knowledge distillation) is in progress — that one will have the complete recipe.