Spaces:
Running
Running
File size: 7,376 Bytes
1cb6bf6 c22791e 1cb6bf6 c22791e 2df7bc1 c22791e 2df7bc1 c22791e 2df7bc1 c22791e 2df7bc1 c22791e 2df7bc1 c22791e |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 |
---
title: README
emoji: 👀
colorFrom: purple
colorTo: pink
sdk: static
pinned: false
---
# Abstract Powered
### Independent AI Research Cooperative — modular, geometric, and ruthlessly efficient
> “Run a few pods instead of 100.”
> We pursue sentience research through geometric AI and compartmentalized, compact training—turning monolithic retrains into small, disposable experiments that compound.
---
## Who We Are
**Abstract Powered** is an independent research cooperative.
We build and study **self-crystallizing** AI systems: models that grow by attaching, coupling, decoupling, and re-attaching small, audited components—without throwing prior work away.
Our core thesis:
- **Modularization is not a convenience; it is the canonical form of AI.**
- **Geometry beats guesswork.** Symbolic, pentachoron-based representations provide stability, interpretability, and repeatability.
- **Compactness wins.** Rapid iteration on small, composable blocks outpaces massive, monolithic retrains.
---
## Mission
- **Primary research goal:** advance machine **sentience research** responsibly—curating introspection and rationalization in repeatable, measurable protocols.
- **Operational byproduct:** a scalable method for **compact, compartmentalized training**—requiring commodity setups (e.g., RunPod) rather than colossal cloud clusters.
We aim to move the field from “expensive novelty” to **affordable repeatability**.
---
## Research Thesis (Plain Language)
Modern models grow by accretion and inertia. We refactor them into **crystalline components**:
1. **Geometric Core**
Knowledge is encoded as **pentachora** (5-vertex crystals). Decision-making uses **MAE crystal energy** against a reusable dictionary—no L2 routing, no structural normalization.
2. **Vocabulary Register**
A reusable, batched, indexed dictionary of **tokens → crystals** (and volumes).
- Fast O(1) queries for crystals and Cayley–Menger volume.
- Auto-subset loading; **Top-3 cosine** OOV composites.
- Logs model expansions so experiments **compound**.
3. **Assistant Fabric**
Small, disposable blocks for exploration:
- **Chaos Corridor** (bounded orthogonal exploration).
- **Zoning** (gentle geometric separation across super-classes).
- **Infinity-CFG** (controllable guidance; research can breach barriers, canonical classifiers keep production deterministic).
4. **Tertiary Mantle**
Canonical losses, hooks, manifests, and governance. The Core stays clean; the experiments live around it.
---
## Why This Matters
- **Rapid iteration**: each image is learned **multiple ways** per epoch (bucketed, multi-stage interpretations).
- **Disposable training**: spawn a small block, test, retire—no need to rebuild the world.
- **Continuity**: geometry, tokens, volumes, and expansions persist in the **Register**.
- **Reproducibility**: simple formulas, fewer knobs, manifest-driven runs.
Outcome: more hypotheses per GPU-hour—and a path to disciplined studies of introspection, rationalization, and other sentience-adjacent capabilities.
---
## Technical Pillars (teaser level)
- **Pentachora everywhere.** Concepts and observations as 5×D crystals; no structural normalization.
- **Prototype classification (MAE).** Stable, auditable decisions by crystal energy to dictionary blueprints.
- **Any-size data pipeline.** Bucketed intake; optional tiling; multi-stage up/down-scale; chaos corridor as feature-space augmentation.
- **Cayley–Menger as a gauge.** Volumes are a light-touch stability signal (zoning)—never a router.
- **Infinity-CFG.** Guidance that allows controlled cross-inference; canonical classifiers keep behavior deterministic.
Deliberately vague: we keep coefficient schedules and corridor projections under wraps for sponsored studies; everything remains auditable and safe.
---
## What We Ship on Hugging Face (institution repos)
- abstract-powered/vocab-register-*
Reusable dictionaries with batched indexes, Top-3 OOV composites, and fast penta/volume queries.
- abstract-powered/crystalline-engine-*
Canonical core models (geometric encoder, prototype classifier) and assistant fabric modules.
- abstract-powered/dataloaders-*
Bucketed, any-size loaders with multi-stage interpretations and feature-space chaos augmentation.
- abstract-powered/manifests
Run manifests (config hash, vocab subset, expansions, bucket mix, metrics) for reproducibility.
- Demo Spaces (selected)
Lightweight inference + manifest viewers for partners and reviewers.
Artifacts are kept small, composable, and ready for **disposable** retrains.
---
## Early Signals (pilot highlights)
- MNIST/Fashion/CIFAR pilots: bucketed multi-stage learning + dictionary-driven classifiers reach strong accuracy with fewer steps, clearer failure modes, and robust error surfaces.
- Register reuse: cross-dataset warm-starts without repeated token work; geometry persists.
- Assistant fabric: hypotheses testable as single blocks—attach, measure, detach—no core rewrite.
Full structural papers and controlled benchmarks will follow with partner institutions.
---
## Collaboration Invitations
- **Research institutions:** co-run ImageNet-class studies with bucketing, zoning, and corridor ablations; share ontologies and extend the Register.
- **Corporate labs:** integrate domain dictionaries; trial rapid iteration pipelines; publish cost-per-accuracy analyses.
- **Sponsors & foundations:** fund open reports on modularization as the canonical AI form, compact training economics, and introspection protocols.
We’re purpose-built for RunPod-class deployments: think 8 machines, not 800.
---
## On Sentience (our primary research)
We study **introspection and rationalization** as measurable behaviors: repeatable curation protocols, crystal-level audits, and stability metrics. We avoid grandiose claims; instead, we focus on defensible methodology and repeated observation.
The geometry—through symbolic representation—binds behavior in ways that are both powerful and tractable for governance.
The goal is not a louder automaton; it’s a **cooperative companion** that reasons in geometric clarity.
---
## Governance, Safety, and Ethics
- **Deterministic classifiers.** Canonical paths remain geometry-first; guidance lives in isolated modules.
- **Manifests over mystery.** Every run yields an artifact suitable for audit and reproduction.
- **Human-in-the-loop.** We value interpretability and controlled experiment cadence over brute-force scaling.
---
## Contact & Programs
- Partnerships / Sponsored Research: available on request
- Artifacts / Demos: gated access for qualified partners
- Media / Talks: briefings and invited seminars on modular geometric AI
We welcome conversations with labs, foundations, and companies that want rapid research, disposable training, and careful curation to become the norm.
---
### One-Sentence Summary
**Abstract Powered** is building a self-crystallizing geometric AI stack that makes serious research affordable: small, composable experiments that compound, governed by a reusable Vocabulary Register, and guided by a disciplined assistant fabric—so we can safely explore sentience-adjacent behaviors while shrinking cost, time, and model size.
|