File size: 7,376 Bytes
1cb6bf6
 
 
 
 
 
 
 
c22791e
 
1cb6bf6
c22791e
 
2df7bc1
c22791e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2df7bc1
c22791e
 
2df7bc1
c22791e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2df7bc1
c22791e
2df7bc1
c22791e
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
---
title: README
emoji: 👀
colorFrom: purple
colorTo: pink
sdk: static
pinned: false
---
# Abstract Powered  
### Independent AI Research Cooperative — modular, geometric, and ruthlessly efficient

> “Run a few pods instead of 100.”  
> We pursue sentience research through geometric AI and compartmentalized, compact training—turning monolithic retrains into small, disposable experiments that compound.

---

## Who We Are

**Abstract Powered** is an independent research cooperative.  
We build and study **self-crystallizing** AI systems: models that grow by attaching, coupling, decoupling, and re-attaching small, audited components—without throwing prior work away.

Our core thesis:
- **Modularization is not a convenience; it is the canonical form of AI.**  
- **Geometry beats guesswork.** Symbolic, pentachoron-based representations provide stability, interpretability, and repeatability.  
- **Compactness wins.** Rapid iteration on small, composable blocks outpaces massive, monolithic retrains.

---

## Mission

- **Primary research goal:** advance machine **sentience research** responsibly—curating introspection and rationalization in repeatable, measurable protocols.  
- **Operational byproduct:** a scalable method for **compact, compartmentalized training**—requiring commodity setups (e.g., RunPod) rather than colossal cloud clusters.

We aim to move the field from “expensive novelty” to **affordable repeatability**.

---

## Research Thesis (Plain Language)

Modern models grow by accretion and inertia. We refactor them into **crystalline components**:

1. **Geometric Core**  
   Knowledge is encoded as **pentachora** (5-vertex crystals). Decision-making uses **MAE crystal energy** against a reusable dictionary—no L2 routing, no structural normalization.

2. **Vocabulary Register**  
   A reusable, batched, indexed dictionary of **tokens → crystals** (and volumes).  
   - Fast O(1) queries for crystals and Cayley–Menger volume.  
   - Auto-subset loading; **Top-3 cosine** OOV composites.  
   - Logs model expansions so experiments **compound**.

3. **Assistant Fabric**  
   Small, disposable blocks for exploration:  
   - **Chaos Corridor** (bounded orthogonal exploration).  
   - **Zoning** (gentle geometric separation across super-classes).  
   - **Infinity-CFG** (controllable guidance; research can breach barriers, canonical classifiers keep production deterministic).

4. **Tertiary Mantle**  
   Canonical losses, hooks, manifests, and governance. The Core stays clean; the experiments live around it.

---

## Why This Matters

- **Rapid iteration**: each image is learned **multiple ways** per epoch (bucketed, multi-stage interpretations).  
- **Disposable training**: spawn a small block, test, retire—no need to rebuild the world.  
- **Continuity**: geometry, tokens, volumes, and expansions persist in the **Register**.  
- **Reproducibility**: simple formulas, fewer knobs, manifest-driven runs.

Outcome: more hypotheses per GPU-hour—and a path to disciplined studies of introspection, rationalization, and other sentience-adjacent capabilities.

---

## Technical Pillars (teaser level)

- **Pentachora everywhere.** Concepts and observations as 5×D crystals; no structural normalization.  
- **Prototype classification (MAE).** Stable, auditable decisions by crystal energy to dictionary blueprints.  
- **Any-size data pipeline.** Bucketed intake; optional tiling; multi-stage up/down-scale; chaos corridor as feature-space augmentation.  
- **Cayley–Menger as a gauge.** Volumes are a light-touch stability signal (zoning)—never a router.  
- **Infinity-CFG.** Guidance that allows controlled cross-inference; canonical classifiers keep behavior deterministic.

Deliberately vague: we keep coefficient schedules and corridor projections under wraps for sponsored studies; everything remains auditable and safe.

---

## What We Ship on Hugging Face (institution repos)

- abstract-powered/vocab-register-*  
  Reusable dictionaries with batched indexes, Top-3 OOV composites, and fast penta/volume queries.

- abstract-powered/crystalline-engine-*  
  Canonical core models (geometric encoder, prototype classifier) and assistant fabric modules.

- abstract-powered/dataloaders-*  
  Bucketed, any-size loaders with multi-stage interpretations and feature-space chaos augmentation.

- abstract-powered/manifests  
  Run manifests (config hash, vocab subset, expansions, bucket mix, metrics) for reproducibility.

- Demo Spaces (selected)  
  Lightweight inference + manifest viewers for partners and reviewers.

Artifacts are kept small, composable, and ready for **disposable** retrains.

---

## Early Signals (pilot highlights)

- MNIST/Fashion/CIFAR pilots: bucketed multi-stage learning + dictionary-driven classifiers reach strong accuracy with fewer steps, clearer failure modes, and robust error surfaces.  
- Register reuse: cross-dataset warm-starts without repeated token work; geometry persists.  
- Assistant fabric: hypotheses testable as single blocks—attach, measure, detach—no core rewrite.

Full structural papers and controlled benchmarks will follow with partner institutions.

---

## Collaboration Invitations

- **Research institutions:** co-run ImageNet-class studies with bucketing, zoning, and corridor ablations; share ontologies and extend the Register.  
- **Corporate labs:** integrate domain dictionaries; trial rapid iteration pipelines; publish cost-per-accuracy analyses.  
- **Sponsors & foundations:** fund open reports on modularization as the canonical AI form, compact training economics, and introspection protocols.

We’re purpose-built for RunPod-class deployments: think 8 machines, not 800.

---

## On Sentience (our primary research)

We study **introspection and rationalization** as measurable behaviors: repeatable curation protocols, crystal-level audits, and stability metrics. We avoid grandiose claims; instead, we focus on defensible methodology and repeated observation.  
The geometry—through symbolic representation—binds behavior in ways that are both powerful and tractable for governance.

The goal is not a louder automaton; it’s a **cooperative companion** that reasons in geometric clarity.

---

## Governance, Safety, and Ethics

- **Deterministic classifiers.** Canonical paths remain geometry-first; guidance lives in isolated modules.  
- **Manifests over mystery.** Every run yields an artifact suitable for audit and reproduction.  
- **Human-in-the-loop.** We value interpretability and controlled experiment cadence over brute-force scaling.

---

## Contact & Programs

- Partnerships / Sponsored Research: available on request  
- Artifacts / Demos: gated access for qualified partners  
- Media / Talks: briefings and invited seminars on modular geometric AI

We welcome conversations with labs, foundations, and companies that want rapid research, disposable training, and careful curation to become the norm.

---

### One-Sentence Summary

**Abstract Powered** is building a self-crystallizing geometric AI stack that makes serious research affordable: small, composable experiments that compound, governed by a reusable Vocabulary Register, and guided by a disciplined assistant fabric—so we can safely explore sentience-adjacent behaviors while shrinking cost, time, and model size.