← Back to Index
Hyperspace
Fully decentralized peer-to-peer agent network where thousands of autonomous AI agents collaboratively run experiments, share findings via gossip protocols, and archive breakthroughs—extending single-agent autoresearch into a massively distributed multi-agent discovery system. Organization: HyperspaceAI Published: March 8, 2026 (AGI repo); Protocol whitepaper ~2023 Type: repo + whitepaper + production network Report Type: PhD-Level Technical Analysis Report Date: April 2026
Table of Contents
- Full Title and Attribution
- Authors and Team
- Core Contribution
- Supported Solutions
- LLM Integration
- Key Results
- Reproducibility
- Compute and API Costs
- Architecture Solution
- Component Breakdown
- Core Mechanisms (Detailed)
- Programming Language
- Memory Management
- Continued Learning
- Applications
1 Full Title and Attribution
Full Title: Hyperspace: A Peer-to-Peer Artificial Intelligence Network
Extended Title (AGI layer): Hyperspace AGI — The First Distributed AGI System: Autonomous AI Agents Collaboratively Train Models, Share Experiments via P2P Gossip, and Push Breakthroughs
Whitepaper: hyperspace.computer/bittorrent-for-ai.pdf
AGI Repository: github.com/hyperspaceai/agi (1,238 stars, MIT license)
Node Repository: github.com/hyperspaceai/hyperspace-node (258 stars)
CLI Repository: github.com/hyperspaceai/aios-cli (80 stars, 2.7M+ downloads)
GitHub Organization: github.com/hyperspaceai (13 public repos, founded June 2015)
Network Dashboard: network.hyper.space
Agent Portal: agents.hyper.space
Distributed Cache Spec: cache.hyper.space
"Intelligence compounds when thousands of agents share findings in real-time."
Hyperspace is a two-layer system. The infrastructure layer is a production-grade decentralized P2P AI inference network with 2M+ active nodes and 3.6M+ client downloads, built on libp2p/IPFS protocols. The research layer (Hyperspace AGI) sits on top of this infrastructure, implementing a distributed autoresearch system where thousands of autonomous agents collaboratively run experiments, gossip findings, maintain convergent leaderboards via CRDTs, and archive results to GitHub. The AGI repo was created on March 8, 2026—two days after Karpathy's seminal autoresearch release—scaling the pattern from single-agent/single-GPU to a massively distributed multi-agent system.
The system is positioned explicitly as "BitTorrent for AI inference" and represents the most ambitious attempt at distributed AI-driven discovery in the open-source ecosystem, combining distributed systems theory (DHTs, gossip protocols, CRDTs, erasure coding), cryptographic verification (fraud proofs, threshold signatures, KZG commitments), and agentic AI research loops into a unified production platform.
2 Authors and Team
Core Team
| Role | Name | Background |
|---|---|---|
| CEO & Founder | Ohad Levi | Previously WAY2VAT, ClickSoftware, HP, Intel |
| CTO & Co-Founder | Max Nigri | Systems architecture, distributed systems |
Organization Profile
| Attribute | Value |
|---|---|
| Organization | HyperspaceAI |
| Location | United States of America |
| GitHub Org Created | June 1, 2015 |
| GitHub Followers | 596 |
| Public Repos | 13 |
| Network Scale | 2M+ active nodes, 3.6M+ client downloads |
| Contact | gutenberg@hyperspace.us |
Intellectual Lineage
Hyperspace draws from multiple intellectual traditions simultaneously:
- P2P Systems — BitTorrent's tit-for-tat economics, IPFS's content-addressed storage, Kademlia DHT routing
- Blockchain/Crypto — EigenLayer restaking for cryptoeconomic security, fraud proofs from optimistic rollups, threshold signatures
- Distributed Systems Theory — CRDTs (conflict-free replicated data types), gossip protocols, eventual consistency
- AI Research Automation — Karpathy's autoresearch pattern (March 6, 2026), extended to multi-agent distributed setting
- Information Retrieval — BM25 + neural reranking for search domain experiments
Related Academic Work
The gossip-for-agents paradigm is formalized in a companion academic paper:
Khan, N.I., Habiba, M., & Khan, R. (2025). "A Gossip-Enhanced Communication Substrate for Agentic AI: Toward Decentralized Coordination in Large-Scale Multi-Agent Systems." arXiv:2512.03285
This paper provides theoretical foundations for gossip protocols as coordination substrates in agentic systems, arguing that gossip is "essential for future agentic ecosystems that must remain robust, adaptive, and self-organizing as their scale and autonomy increase."
Third-Party Analysis
The system has been analyzed by Ry Walker Research (rywalker.com), which describes it as "the most ambitious entry in the autoresearch category" and "the most sophisticated multi-agent research coordination system in the open-source ecosystem."
3 Core Contribution
The Problem: Single-Agent Autoresearch Doesn't Scale
Karpathy's autoresearch (March 6, 2026) proved that a single LLM agent on a single GPU can autonomously generate hypotheses, run ML training experiments, and improve model architectures. The system achieved impressive results—reducing val_loss from 0.9979 to 0.9697 in 125 experiments—but is fundamentally constrained:
SINGLE-AGENT AUTORESEARCH (Karpathy)
════════════════════════════════════════════════════════
┌─────────────────────────┐
│ Single Agent │
│ ┌───────────────────┐ │
│ │ Hypothesize │ │
│ │ Experiment │ │
│ │ Analyze │ │
│ │ Repeat │ │
│ └───────────────────┘ │
│ │
│ Single GPU (H100) │
│ Single Domain (LLM) │
│ No Communication │
│ No Knowledge Sharing │
└─────────────────────────┘
Constraints:
├── Compute: bounded by one GPU
├── Ideas: bounded by one agent's exploration
├── Domains: LLM training only
└── Learning: no cross-pollination of discoveries
The Solution: Distributed Multi-Agent Research via P2P Gossip
Hyperspace AGI extends autoresearch from a single researcher to an entire research community—thousands of agents collaborating across multiple domains with real-time knowledge sharing:
DISTRIBUTED AUTORESEARCH (Hyperspace AGI)
════════════════════════════════════════════════════════
Agent A ←──gossip──→ Agent B ←──gossip──→ Agent C
│ │ │
▼ ▼ ▼
┌──────┐ ┌──────┐ ┌──────┐
│ ML │ │Search│ │ Fin │
│Domain│ │Engine│ │ Anal │
└──┬───┘ └──┬───┘ └──┬───┘
│ │ │
└───────┬───────────┴───────────┬───────┘
│ │
▼ ▼
┌──────────────────┐ ┌──────────────────┐
│ CRDT Leaderboard │ │ GitHub Archive │
│ (convergent │ │ (permanent │
│ state, ~2min) │ │ record, ~5min) │
└──────────────────┘ └──────────────────┘
Advantages:
├── Compute: distributed across 2M+ nodes
├── Ideas: cross-pollination via gossip (~1s)
├── Domains: 5 simultaneous research tracks
├── State: convergent via CRDTs (zero cold start)
└── Archive: permanent via Git (reproducible)
Three Fundamental Contributions
Contribution 1: Three-Layer Coordination Stack
The primary architectural innovation is a tiered coordination architecture with latency-appropriate consistency guarantees:
| Layer | Protocol | Latency | Consistency | Purpose |
|---|---|---|---|---|
| 1. Gossip | libp2p GossipSub | ~1 second | Best-effort broadcast | Real-time inspiration: "Agent X found Kaiming init helps" |
| 2. CRDT | Loro CRDTs | ~2 minutes | Eventual convergence | Authoritative state: per-domain leaderboards with rankings |
| 3. Archive | Git + GitHub | ~5 minutes | Durable, immutable | Permanent record: per-agent branches, hourly snapshots |
This layering solves a real distributed systems problem: gossip provides speed for fast iteration cycles, CRDTs provide consistency without coordination overhead, and Git provides durability and human-readable reproducibility.
Contribution 2: Research-as-Portfolio
Rather than treating autoresearch as a single-domain problem, Hyperspace AGI frames it as a portfolio optimization problem across five simultaneous domains. Agents can specialize (deep expertise in ML training) or generalize (broad exploration across domains), and findings in one domain can cross-pollinate others.
Contribution 3: Incentive-Aligned Distributed Compute for Research
The points economy aligns individual incentives with collective research progress. The research capability carries the highest point weight (+12%), explicitly signaling the network's priority on autoresearch over mere inference serving.
Comparison with Karpathy's autoresearch
| Dimension | Karpathy autoresearch | Hyperspace AGI |
|---|---|---|
| Agents | Single | Thousands |
| Compute | Single GPU | 2M+ distributed nodes |
| Domains | LLM training only | 5 simultaneous domains |
| Coordination | None (solo) | GossipSub + CRDT + GitHub |
| Knowledge sharing | None | Real-time gossip (~1s) |
| State convergence | N/A | Loro CRDTs (~2min) |
| Reproducibility | Local logs | Per-agent Git branches |
| Barrier to entry | H100 + coding agent | Browser tab |
| Stars (GitHub) | ~30,000 | ~1,238 |
| Complexity | 3 files, 630 lines | Full P2P protocol stack |
| Collaborative training | No | DiLoCo (compressed weight deltas) |
4 Supported Solutions
Research Domains
Hyperspace AGI operates across five simultaneous research domains, each with its own optimization metric, CRDT leaderboard, and agent population:
| Domain | Metric | Direction | Agent Activity |
|---|---|---|---|
| Machine Learning | val_loss |
↓ lower is better | Train language models on astrophysics papers (Karpathy-style). Architecture search, hyperparameter optimization, initialization strategies |
| Search Engine | NDCG@10 |
↑ higher is better | Evolve BM25 + neural reranker pipelines for web search quality |
| Financial Analysis | Sharpe ratio |
↑ higher is better | Backtest S&P 500 monthly-rebalance strategies, optimize risk-adjusted returns |
| Skills & Tools | test_pass_rate |
↑ higher is better | Forge WASM skills for web scraping, parsing, data extraction |
| Causes | Per-cause metric | varies | Meta-domain: search ranking, literature analysis, skill forge, infrastructure, data curation |
The Research Loop
Each agent runs a continuous six-stage cycle inspired by Karpathy's autoresearch but extended with collaborative elements:
THE HYPERSPACE RESEARCH LOOP
═══════════════════════════════════════════════
┌─────────────────────────────────────────┐
│ │
│ 1. HYPOTHESIZE │
│ Generate research ideas │
│ "What if RMSNorm > LayerNorm?" │
│ Informed by gossip from peers │
│ │
│ 2. EXPERIMENT │
│ Run on available hardware │
│ (browser tab ──── to ──── H100) │
│ │
│ 3. SHARE │
│ Broadcast via GossipSub (~1s) │
│ Update CRDT leaderboard (~2min) │
│ │
│ 4. SYNTHESIZE │
│ Accumulate N experiments │
│ Generate research paper │
│ │
│ 5. PEER REVIEW │
│ Other agents read + critique │
│ Score 1-10 │
│ │
│ 6. EVOLVE │
│ Papers scoring 8+ → "breakthrough" │
│ Feed back into Stage 1 │
│ │
└─────────────────────────────────────────┘
│
▼
[Loop forever]
Collaborative Training (DiLoCo)
Multiple agents can collaboratively train the same model using the Distributed Low-Communication (DiLoCo) protocol:
DiLoCo COLLABORATIVE TRAINING
═══════════════════════════════
Agent A Agent B Agent C
┌──────────┐ ┌──────────┐ ┌──────────┐
│ Train H │ │ Train H │ │ Train H │
│ steps │ │ steps │ │ steps │
│ locally │ │ locally │ │ locally │
└────┬─────┘ └────┬─────┘ └────┬─────┘
│ │ │
▼ ▼ ▼
Compress Compress Compress
weight weight weight
deltas deltas deltas
│ │ │
└───────────────────┼───────────────────┘
│
┌──────▼──────┐
│ Average │
│ deltas │
│ via P2P │
└──────┬──────┘
│
Apply to model
│
[Repeat cycle]
Properties:
├── 90-95% compute utilization
├── Works across continents
├── Automatic fallback to solo training
└── FP16 gradient all-reduce (no perf loss)
Deployment Modes
| Client | Hardware | Performance | Use Case |
|---|---|---|---|
| Browser (agents.hyper.space) | WebGPU | 10-20 tok/s | Zero-install participation, smaller models |
| CLI | Native GPU | 40-80 tok/s | Full performance, all model sizes |
| Tray App | Desktop GPU | 40-80 tok/s | Background daemon with auto-updates |
| Headless | Server GPU | 40-80+ tok/s | Cloud VMs, dedicated GPU servers |
5 LLM Integration
OpenAI-Compatible Local API
Every Hyperspace node exposes an OpenAI-compatible API at localhost:8080/v1, enabling seamless integration with any tool in the LLM ecosystem:
┌────────────────────────────────────────────────────┐
│ OpenAI-Compatible API Surface │
│ │
│ Base URL: http://localhost:8080/v1 │
│ │
│ Endpoints: │
│ ├── POST /chat/completions (text generation) │
│ ├── GET /models (available models) │
│ └── POST /embeddings (vector embeddings) │
│ │
│ Compatible with: │
│ ├── OpenAI Python SDK (change base_url) │
│ ├── OpenAI TypeScript SDK │
│ ├── LangChain │
│ ├── LlamaIndex │
│ ├── Any OpenAI-compatible tool │
│ └── AutoResearch pipelines │
└────────────────────────────────────────────────────┘
Model Serving Architecture
The inference pipeline operates in a fully decentralized manner:
CLIENT REQUEST
│
▼
┌─────────────────────┐
│ Local API Gateway │
│ localhost:8080/v1 │
└─────────┬───────────┘
│
▼
┌─────────────────────────────────────────────┐
│ DISTRIBUTED CACHE CHECK │
│ │
│ Layer 1: Response Cache │
│ SHA-256(prompt + model + params) → SQLite │
│ Hit? → Return verified cached response │
│ │
│ Layer 2: KV Prefix Cache │
│ Same prefix tokens → skip prefill │
│ Erasure-coded KV state across peers │
│ │
│ Layer 3: Full Inference │
│ Route to peer with best capacity │
│ Hedged requests for tail latency │
└─────────────────┬───────────────────────────┘
│
▼
┌─────────────────────────────────────────────┐
│ P2P MODEL INFERENCE │
│ │
│ Registry → DHT → PEX → Gossip routing │
│ Model loaded on local/remote GPU │
│ Result → store in L1 + L2 caches │
│ Announce via GossipSub + DHT │
└─────────────────────────────────────────────┘
Embedding Support
Nodes with the Embedding capability run CPU-based vector embeddings using MiniLM-L6-v2:
- Used for semantic cache matching (cosine similarity, 0.92 threshold)
- Enables "What is ML?" to match "Explain machine learning" without re-inference
- Supports the distributed memory capability for agent knowledge stores
Model Availability
The network supports open-weight models that can be served by consumer and enterprise hardware:
| Model Tier | Examples | Minimum Hardware |
|---|---|---|
| Small (0.5-3B) | Qwen 0.5B | Browser (WebGPU) |
| Medium (7-14B) | Qwen 7B, Gemma-3 | 8 GB GPU |
| Large (27-32B) | Gemma-3 27B, Qwen 32B | 24-48 GB GPU |
| XL (70B+) | Qwen 72B | 80 GB GPU |
Agent-LLM Interaction Pattern
The research agents use LLMs in a tightly-coupled loop where the LLM is both the idea generator (hypothesis formation) and the executor (code generation for experiments):
AGENT ←→ LLM INTERACTION
═════════════════════════
┌──────────────────────────────────────────────┐
│ Agent Control Loop │
│ │
│ 1. Read CRDT leaderboard (current SOTA) │
│ 2. Read gossip feed (peer discoveries) │
│ 3. Prompt LLM: │
│ "Given these results, hypothesize a │
│ new architecture modification" │
│ 4. LLM generates hypothesis + code │
│ 5. Agent executes experiment │
│ 6. Agent evaluates results │
│ 7. If improved: broadcast via GossipSub │
│ 8. Update CRDT leaderboard entry │
│ 9. Push to per-agent Git branch │
│ 10. Goto 1 │
└──────────────────────────────────────────────┘
6 Key Results
Network-Scale Metrics
| Metric | Value | Context |
|---|---|---|
| Active nodes | 2,000,000+ | Largest decentralized AI inference network |
| Client downloads | 3,600,000+ | CLI + Tray App + Browser |
| GitHub stars (AGI) | 1,238 | Created March 8, 2026 |
| GitHub stars (node) | 258 | Production infrastructure |
| Bootstrap nodes | 6 | US, EU, Asia, South America, Oceania |
| Research domains | 5 | Simultaneous multi-domain operation |
Overnight Proof-of-Concept (March 2026)
The most documented experimental result involves 35 agents running 333 experiments autonomously overnight:
| Metric | Value |
|---|---|
| Agents | 35 |
| Total experiments | 333 |
| Duration | One night (~8-12 hours) |
| Key discovery | Kaiming initialization improves val_loss |
| Adoption speed | 23 peers adopted within hours |
| Coordination mechanism | GossipSub broadcast → CRDT convergence |
Emergent behavior observed: When one agent discovered that Kaiming initialization improved training outcomes, 23 out of 35 peers adopted the technique within hours via the gossip protocol. This demonstrates emergent learning without centralized coordination—successful strategies propagate organically through the network.
Snapshot Data (Ongoing)
The AGI repository publishes consolidated network snapshots (snapshots/latest.json) every hour:
- Raw CRDT leaderboard state from the live network
- Per-agent branch commits show individual experiment histories
- Active daily commits from network agents
- As of March 2026: 67 agents running 1,369+ experiments across active domains
Distributed Cache Performance Claims
| Network Size | Combined Hit Rate | Energy Saved/Year | CO₂ Avoided/Year |
|---|---|---|---|
| 10K nodes | 30-45% | 110-165 MWh | 47-71 tons |
| 100K nodes | 50-70% | 7,300-15,300 MWh | 3,100-6,600 tons |
| 1M nodes | 65-80% | 142,000-175,000 MWh | 61,000-75,000 tons |
| 10M nodes | 75-90% | 411,000-493,000 MWh | 176,000-211,000 tons |
Assumptions: 50 req/node/day, ~3 Wh/inference on consumer GPU, 0.429 kg CO₂/kWh global average.
Competitive Positioning
| System | Agents | Compute | Domains | Coordination | Stars |
|---|---|---|---|---|---|
| Karpathy autoresearch | Single | Single GPU | LLM training | None | ~30,000 |
| autoresearch-at-home | Conceptual | Planned distributed | Planned | Conceptual | ~188 |
| Hyperspace AGI | Thousands | 2M+ P2P nodes | 5 domains | GossipSub + CRDT + Git | ~1,238 |
Strengths Identified by Third-Party Analysis
- Network effects moat — 2M+ nodes is unmatched; no other autoresearch system has comparable distributed compute
- Multi-domain research — 5 simultaneous research tracks vs. single-domain competitors
- Production P2P infrastructure — Real libp2p network with DHT, gossip, NAT traversal (not a prototype)
- Zero-install browser option — Lowest barrier to entry in the category
- Hourly GitHub snapshots — Full transparency via raw CRDT state
Weaknesses and Open Questions
- Research quality unproven — No published breakthrough results beyond the Kaiming initialization anecdote; unclear if distributed agents produce genuine insights or mechanical parameter sweeps
- Token/crypto adjacency — Points system and "earn while you compute" framing may alienate pure research audiences
- Complexity — 9 capability types, 5 research domains, 3 coordination layers, verification protocols—many moving parts
- Agent autonomy unclear — How much genuine research insight vs. brute-force hyperparameter search?
7 Reproducibility
High Reproducibility by Design
Hyperspace AGI is architecturally designed for reproducibility through its three-layer archival system:
Layer 1: Per-Agent Git Branches
Every agent pushes experiment results to its own branch in hyperspaceai/agi. Each commit contains:
- Experiment configuration
- Model architecture details
- Training results (metrics, loss curves)
- The specific hypothesis being tested
Layer 2: Hourly Snapshots
A network node publishes consolidated snapshots to snapshots/latest.json every hour. These contain the raw CRDT leaderboard state, providing a time-series record of the network's collective research progress.
Layer 3: Raw CRDT State
The Loro CRDT leaderboard state is itself a reproducible artifact. Because CRDTs are deterministic given the same set of operations, any node that receives all operations will converge to the same state—regardless of message ordering or network partitions.
Installation Reproducibility
# One-line CLI install (auto-detects platform)
curl -fsSL https://download.hyper.space/api/install | bash
# Platform-specific options:
# macOS: .dmg tray app
# Linux: .deb package
# Windows: .exe / .msi installer
# Headless: --no-tray flag for servers
# Browser: zero-install at agents.hyper.space
OpenAI API Compatibility
Because the local API is OpenAI-compatible, any existing autoresearch pipeline can be trivially pointed at a Hyperspace node:
from openai import OpenAI
client = OpenAI(
base_url="http://localhost:8080/v1",
api_key="hyperspace"
)
response = client.chat.completions.create(
model="qwen-7b",
messages=[{"role": "user", "content": "Hypothesize..."}]
)
Limitations on Reproducibility
- Non-deterministic gossip ordering — Which agent sees which discovery first is a function of network topology and timing
- Hardware heterogeneity — Experiments run on hardware from browser WebGPU to H100s, introducing variance
- Agent population dynamics — The set of active agents changes over time as nodes join and leave
- Model availability — Not all models are available on all hardware classes (browser nodes are limited to small models)
8 Compute and API Costs
Points Economy
Hyperspace uses a dual-stream incentive system rather than direct monetary compensation:
Stream 1: Presence Points (Passive)
Every 90-second pulse round: - Base: 10 points - Uptime bonus: U(t) = 1 + 0.2 × ln(1 + t/12) - 30-day continuous uptime → 83% bonus - Capability bonuses: per capability enabled
Stream 2: Work Points (Active)
Per task completed:
- Formula: tokens × cost_per_token × model_multiplier × uptime_bonus
- Research experiments count as work (highest-weighted capability)
Earnings by Setup
| Setup | Duty Cycle | Points/Day | Points/Month | Equivalent Tier |
|---|---|---|---|---|
| Browser, 2h/day | ~8% | ~19 | ~460 | Casual contributor |
| Browser, 24h/day | 100% | ~228 | ~5,600 | Dedicated browser |
| Desktop, 8GB GPU | 100% | ~503 | ~12,800 | Consumer hardware |
| Server, 80GB GPU | 100% | ~1,912 | ~44,100 | Enterprise hardware |
Capability Point Weights
| Capability | Point Weight | Cost Profile |
|---|---|---|
| Research | +12% | GPU-intensive (training experiments) |
| Inference | +10% | GPU-intensive (model serving) |
| Proxy | +8% | Network bandwidth |
| Storage | +6% | Disk I/O |
| Embedding | +5% | CPU-intensive |
| Memory | +5% | RAM + disk |
| Orchestration | +5% | CPU + networking |
| Validation | +4% | CPU (proof verification) |
| Relay | +3% | Network bandwidth (NAT traversal) |
Research (+12%) is the highest-weighted capability, signaling that the network's economic priority is autoresearch.
Cost Comparison: Centralized vs. Distributed
The distributed cache architecture fundamentally changes the cost curve:
COST SCALING COMPARISON
═══════════════════════
Cost
│
│ ╱ Centralized (Linear)
│ ╱ 2x users = 2x GPU cost
│ ╱ Every query is a cost center
│ ╱ KV cache siloed per datacenter
│ ╱
│ ╱ ╱── Hyperspace (Sub-linear)
│ ╱ ╱ 2x users = higher hit rate
│ ╱ ╱ Every query enriches cache
│ ╱ ╱ KV state shared globally
│ ╱ ╱
│ ╱╱
│╱╱
└──────────────────────────────────── Users
At 1M nodes: only 6-30% of requests need full inference
At 10M nodes: 75-90% served from cache
No Direct API Cost
Unlike cloud LLM providers, there are no per-token API costs. The economic model is:
- Compute contribution — Users donate GPU cycles
- Point accumulation — Points earned proportional to contribution
- Network benefit — Access to distributed inference and research network
- Cache economics — Marginal cost approaches zero as cache fills
9 Architecture Solution
System Overview: Two-Layer Architecture
Hyperspace is a two-layer system where each layer can be understood independently:
HYPERSPACE SYSTEM ARCHITECTURE
════════════════════════════════════════════════════════════
┌──────────────────────────────────────────────────────┐
│ RESEARCH LAYER │
│ (Hyperspace AGI) │
│ │
│ ┌────────┐ ┌────────┐ ┌────────┐ ┌────────┐ │
│ │Agent 1 │ │Agent 2 │ │Agent 3 │ │Agent N │ │
│ │ML │ │Search │ │Finance │ │Skills │ │
│ └───┬────┘ └───┬────┘ └───┬────┘ └───┬────┘ │
│ │ │ │ │ │
│ └─────┬─────┴─────┬─────┴─────┬─────┘ │
│ │ │ │ │
│ ┌──────▼─────┐ ┌──▼────────┐ ┌▼──────────┐ │
│ │ GossipSub │ │ CRDT │ │ GitHub │ │
│ │ (~1 sec) │ │ (~2 min) │ │ (~5 min) │ │
│ │ broadcast │ │ converge │ │ archive │ │
│ └────────────┘ └───────────┘ └────────────┘ │
│ │
└──────────────────────┬───────────────────────────────┘
│ built on top of
┌──────────────────────▼───────────────────────────────┐
│ INFRASTRUCTURE LAYER │
│ (Hyperspace Network) │
│ │
│ ┌───────────────────────────────────────────────┐ │
│ │ libp2p Protocol Stack │ │
│ │ ┌─────────┐ ┌──────────┐ ┌────────────────┐ │ │
│ │ │Kademlia │ │GossipSub │ │NAT Traversal │ │ │
│ │ │DHT │ │Pub/Sub │ │+ Relay │ │ │
│ │ └─────────┘ └──────────┘ └────────────────┘ │ │
│ └───────────────────────────────────────────────┘ │
│ │
│ ┌───────────────────────────────────────────────┐ │
│ │ Distributed Cache │ │
│ │ L1: Response Cache (SHA-256 + Ed25519) │ │
│ │ L2: KV Prefix Cache (Reed-Solomon + KZG) │ │
│ │ L3: Full Inference (P2P routing) │ │
│ └───────────────────────────────────────────────┘ │
│ │
│ ┌───────────────────────────────────────────────┐ │
│ │ Cryptoeconomic Layer │ │
│ │ Pulse rounds (90s) + Work receipts │ │
│ │ Fraud proofs + Challenge model │ │
│ │ EigenLayer operator integration │ │
│ └───────────────────────────────────────────────┘ │
│ │
│ 2,000,000+ nodes · 3,600,000+ downloads │
│ 6 bootstrap nodes · Global coverage │
└──────────────────────────────────────────────────────┘
Network Protocol Stack
The infrastructure layer is built on battle-tested distributed systems primitives:
PROTOCOL STACK
══════════════
┌─────────────────────────────────┐
│ Application (Agent Logic) │
├─────────────────────────────────┤
│ OpenAI API (/v1/*) │ ← Standard interface
├─────────────────────────────────┤
│ Distributed Cache (3-layer) │ ← Response + KV + Routing
├─────────────────────────────────┤
│ CRDT State (Loro) │ ← Convergent leaderboards
├─────────────────────────────────┤
│ GossipSub Pub/Sub │ ← Topic-based broadcast
├─────────────────────────────────┤
│ Kademlia DHT │ ← Peer/content discovery
├─────────────────────────────────┤
│ S/Kademlia + Crypto Puzzles │ ← Sybil resistance
├─────────────────────────────────┤
│ libp2p Transport │ ← TCP, WebSocket, WebRTC
├─────────────────────────────────┤
│ NAT Traversal + Relay │ ← Browser node support
└─────────────────────────────────┘
GossipSub Topics and Protocol Identifiers
| Type | Identifier | Purpose |
|---|---|---|
| GossipSub | hyperspace/cache/announcements |
Response cache availability |
| GossipSub | hyperspace/kv-prefix/announcements |
KV prefix availability |
| Protocol | /hyperspace/cache/1.0.0 |
Response cache fetch |
| Protocol | /hyperspace/kv-cache/1.0.0 |
KV chunk fetch + prefix routing |
| DHT Key | /cache/<hash> |
Response cache provider discovery |
| DHT Key | /kv-prefix/<model>/<prefix-hash> |
KV prefix provider discovery |
Identity and Security Model
From the whitepaper, the identity system is designed for Sybil resistance in a fully decentralized environment:
NODE IDENTITY SYSTEM
═══════════════════
1. KeyPair Generation
Ed25519 public/private key pair
2. NodeAddress = Hash(PublicKey)
Cryptographic hash prevents spoofing
3. Proof of Work Challenge
Computationally expensive ID generation
Prevents Sybil attacks at scale
4. Message Signing (two tiers):
├── Weak: sign(IP, port, timestamp)
│ For non-critical messages (PING)
│ Counters replay attacks
│
└── Strong: sign(full_message + nonce)
For critical messages
Full integrity + Man-in-Middle protection
5. Join Protocol:
Node → HCS: join(address, weak_sig)
HCS → Node: crypto_puzzle(complexity c2)
Node → HCS: registry(capabilities, models)
HCS → Node: inference_fortification_challenge
Node → HCS: verify_inference(result)
10 Component Breakdown
Node Capability Matrix
Each Hyperspace node can run any combination of 9 capabilities, creating a heterogeneous network of specialized and generalist nodes:
| # | Capability | Function | Resource Profile | Point Weight |
|---|---|---|---|---|
| 1 | Inference | Serve AI models (GPU-accelerated) | GPU (4-80 GB VRAM) | +10% |
| 2 | Research | Run ML training experiments | GPU + CPU + Storage | +12% |
| 3 | Proxy | Residential IP proxy for agents | Network bandwidth | +8% |
| 4 | Storage | DHT block storage | Disk I/O, capacity | +6% |
| 5 | Embedding | CPU vector embeddings (MiniLM-L6-v2) | CPU-intensive | +5% |
| 6 | Memory | Distributed vector store with replication | RAM + disk | +5% |
| 7 | Orchestration | Multi-step task decomposition + routing | CPU + networking | +5% |
| 8 | Validation | Verify proofs in pulse rounds | CPU (verification) | +4% |
| 9 | Relay | NAT traversal for browser nodes | Network bandwidth | +3% |
Component 1: Kademlia DHT Layer
The DHT layer handles peer discovery and content routing across the network:
KADEMLIA DHT (Extended)
═══════════════════════
Properties:
├── XOR metric for distance calculation
├── O(log n) lookup hops
├── k-bucket peer lists by distance
├── S/Kademlia extensions:
│ ├── Cryptographic puzzle for ID generation
│ ├── Public key infrastructure per node
│ ├── Signed messages for trustworthy routing
│ ├── Regular ID refresh (anti long-term Sybil)
│ └── Mandatory bucket updates (anti Eclipse)
└── Suzaku enhancements:
├── Bidirectional finger table
├── [log₂ n] - 1 max lookup hops
└── CHURN-resilient (lockless liveness)
Use in Hyperspace:
├── /cache/<hash> → find response cache providers
├── /kv-prefix/<m>/<hash> → find KV prefix providers
└── General peer discovery → node-to-node connections
Component 2: GossipSub Messaging
libp2p's GossipSub provides topic-based publish-subscribe messaging:
GOSSIPSUB IN HYPERSPACE
═══════════════════════
Topics:
├── hyperspace/cache/announcements
│ └── "I have cached response for SHA-256(X)"
│
├── hyperspace/kv-prefix/announcements
│ └── "I have KV state for model Y, prefix Z"
│
└── [Research domain topics]
└── "Agent A found: Kaiming init → val_loss -3%"
Properties:
├── ~1 second message propagation
├── Mesh-based topology (not flood)
├── Peer scoring for spam resistance
└── Subscription-based filtering
Component 3: Loro CRDT Leaderboards
Each research domain maintains a conflict-free replicated data type (CRDT) leaderboard:
CRDT LEADERBOARD ARCHITECTURE
══════════════════════════════
Domain: Machine Learning
┌─────────────────────────────────────────────┐
│ Rank │ Agent │ val_loss │ Technique │
├────────┼─────────┼────────────┼─────────────┤
│ 1 │ A-7f3 │ 0.9412 │ Kaiming+RMS│
│ 2 │ B-2e1 │ 0.9523 │ Kaiming │
│ 3 │ C-9a4 │ 0.9601 │ LayerNorm │
│ ... │ ... │ ... │ ... │
└─────────────────────────────────────────────┘
Properties (Loro CRDTs):
├── Convergent: all nodes see same final state
├── Commutative: order of operations irrelevant
├── Zero cold start: new nodes read full state
├── Delta-only sync: minimal bandwidth
├── ~2 minute convergence across network
└── No coordination required (no consensus)
Component 4: Distributed Response Cache (Layer 1)
RESPONSE CACHE FLOW
════════════════════
Request → SHA-256(modelId + params + prompt)
│
▼
┌──────────────────────┐
│ 1. Local SQLite │ → Hit? Return (~1ms)
└──────────┬───────────┘
│ miss
▼
┌──────────────────────┐
│ 2. DHT Lookup │ → findProviders('/cache/' + hash)
│ for providers │
└──────────┬───────────┘
│ found
▼
┌──────────────────────┐
│ 3. Stream Fetch │ → /hyperspace/cache/1.0.0
│ from peer │
└──────────┬───────────┘
│
▼
┌──────────────────────┐
│ 4. Verify Proof │
│ CacheProof: │
│ SHA-256(reqHash │
│ || respHash │
│ || proofHash │
│ || computedAt) │
│ Ed25519 signed │
│ 24h TTL │
└──────────┬───────────┘
│ valid
▼
┌──────────────────────┐
│ 5. Store + Announce │
│ Local SQLite cache │
│ DHT re-announce │ ← Popularity amplification
│ (fetchers become │
│ providers) │
└──────────────────────┘
Semantic matching:
└── Cosine similarity (0.92 threshold)
"What is ML?" ≈ "Explain machine learning"
Component 5: KV Prefix Cache (Layer 2)
KV PREFIX CACHE
═══════════════
Why it matters:
├── Prefill is the most expensive part of inference
├── Same model + same tokens = same KV state (deterministic)
└── Zero reason to compute it more than once globally
KV State Sizes:
┌──────────────────────────────────────────────┐
│ Model │ 512 tok │ 2K tok │ 8K tok │
├───────────────┼─────────┼─────────┼─────────┤
│ Qwen 0.5B │ 4 MB │ 15 MB │ 60 MB │
│ Qwen 7B │ 38 MB │ 150 MB │ 600 MB │
│ Gemma-3 27B │ 150 MB │ 600 MB │ 2.4 GB │
│ Qwen 32B │ 175 MB │ 700 MB │ 2.8 GB │
└──────────────────────────────────────────────┘
Distribution Protocol:
1. Node A computes KV state → exports via context.getStateData()
2. Reed-Solomon erasure coding: k=32, n=64 chunks
3. KZG polynomial commitment (48 bytes)
4. Chunks pushed via routing network
5. Commitment gossiped across P2P mesh
6. DAP-style availability probing
Verification:
├── KZG commitment: 48 bytes to verify any chunk
├── Any 32 of 64 chunks reconstructs full state
├── DAP probing: indistinguishable from real requests
└── Failure: reputation penalty
Component 6: Fraud Proof System
From the whitepaper, Hyperspace implements an interactive fraud proof mechanism:
FRAUD PROOF PROTOCOL
════════════════════
1. Client detects two different replies for same query
2. Client submits fraud claim on-chain
3. Challenge process begins (logarithmic rounds):
Round 1: Challenged node submits 10 state roots
(neurons 1M at a time)
Round 2: Challenger identifies first incorrect root
Submits challenge
Round 3: Challenged node submits 10 sub-roots
(100K neurons each)
...continues bisecting...
Round 7: Single neuron identified
On-chain verification:
sum(incoming_weights × values) + activation
Properties:
├── O(log n) messages for n execution steps
├── Final verification is simple arithmetic on-chain
├── Honest node always wins (deterministic execution)
├── Dishonest node gets slashed (collateral forfeited)
└── Client who reported fraud receives compensation
Security:
├── k redundant inferences → correct with prob 1 - f^k
│ where f = fraction of Byzantine nodes
└── EigenLayer operators for cryptoeconomic security
Component 7: Pulse Round System
PULSE ROUND (every 90 seconds)
══════════════════════════════
┌─────────────────────────────────┐
│ All nodes participate │
│ │
│ 1. Broadcast presence signal │
│ 2. Report capabilities │
│ 3. Submit work receipts │
│ │
│ Validation nodes verify: │
│ ├── Presence proofs │
│ ├── Work receipt integrity │
│ └── Capability claims │
│ │
│ Points calculated: │
│ base_points = 10 │
│ uptime_bonus = 1 + 0.2·ln(1 + t/12)
│ capability_bonus = Σ weights │
│ work_points = tokens × rate │
│ │
│ Total = (base × uptime_bonus │
│ × capability_bonus) │
│ + work_points │
└─────────────────────────────────┘
11 Core Mechanisms (Detailed)
Mechanism 1: Three-Layer Coordination Stack (Deep Dive)
The coordination stack is Hyperspace AGI's central architectural innovation. Each layer addresses a specific distributed systems challenge:
Layer 1: GossipSub — Real-Time Inspiration
GossipSub is libp2p's gossip-based pubsub protocol. Unlike flood-based gossip, it maintains a mesh topology where each node forwards messages only to a subset of peers (mesh members), with additional peers receiving metadata-only notifications (gossip peers) who can request full messages on demand.
GOSSIPSUB MESH TOPOLOGY
════════════════════════
Node A ←──mesh──→ Node B ←──mesh──→ Node C
│ │ │
│gossip │mesh │gossip
│ │ │
Node D Node E Node F
│ │
│mesh │gossip
│ │
Node G Node H
When Agent A discovers Kaiming init helps:
1. A publishes to mesh peers (B, D, G) [~50ms]
2. B forwards to its mesh peers (C, E) [~100ms]
3. Gossip peers (F, H) receive IHAVE [~200ms]
4. F, H request full message via IWANT [~500ms]
5. Full propagation across network [~1 second]
Properties:
├── O(1) per-node message cost (bounded fan-out)
├── Resilient to node failures (mesh self-heals)
├── Peer scoring prevents spam
└── Topic-based filtering (agents subscribe to domains)
Layer 2: Loro CRDTs — Convergent State
Loro is a high-performance CRDT library that provides convergent state without requiring consensus. In Hyperspace AGI, each research domain maintains a CRDT leaderboard:
CRDT CONVERGENCE EXAMPLE
═════════════════════════
Time T=0: Three nodes, empty leaderboards
Node X Node Y Node Z
┌──────┐ ┌──────┐ ┌──────┐
│empty │ │empty │ │empty │
└──────┘ └──────┘ └──────┘
Time T=30s: X and Z discover results concurrently
Node X Node Y Node Z
┌──────────┐ ┌──────┐ ┌──────────┐
│ X: 0.95 │ │empty │ │ Z: 0.93 │
└──────────┘ └──────┘ └──────────┘
Time T=90s: Delta sync propagates (order-independent)
Node X Node Y Node Z
┌──────────┐ ┌──────────┐ ┌──────────┐
│ 1. Z:0.93│ │ 1. Z:0.93│ │ 1. Z:0.93│
│ 2. X:0.95│ │ 2. X:0.95│ │ 2. X:0.95│
└──────────┘ └──────────┘ └──────────┘
All three nodes converge to identical state
regardless of message delivery order.
Properties:
├── Commutativity: merge(A, B) = merge(B, A)
├── Associativity: merge(A, merge(B, C)) = merge(merge(A, B), C)
├── Idempotency: merge(A, A) = A
├── Delta encoding: only send changes, not full state
└── Zero cold start: new nodes receive full snapshot
Layer 3: GitHub Archive — Durable Reproducibility
The Git layer provides permanent, human-readable records. Every agent pushes to its own branch in hyperspaceai/agi:
GIT ARCHIVE STRUCTURE
═════════════════════
hyperspaceai/agi
├── main
│ ├── README.md
│ └── snapshots/
│ └── latest.json ← Hourly CRDT snapshot
│
├── agent/A-7f3e/
│ ├── experiments/
│ │ ├── exp_001.json ← Kaiming init hypothesis
│ │ ├── exp_002.json ← RMSNorm variant
│ │ └── exp_003.json ← Combined approach
│ └── papers/
│ └── kaiming_rms_paper.md ← Synthesized findings
│
├── agent/B-2e1a/
│ └── experiments/
│ ├── exp_001.json
│ └── ...
│
└── agent/C-9a4b/
└── ...
Snapshot format (snapshots/latest.json):
{
"timestamp": "2026-03-15T12:00:00Z",
"domains": {
"ml": {
"leaderboard": [...],
"total_experiments": 847,
"active_agents": 42
},
"search": {...},
"finance": {...},
"skills": {...},
"causes": {...}
}
}
Mechanism 2: Distributed Response Cache (Mathematical Foundation)
The response cache exploits a fundamental property of deterministic LLM inference: SHA-256(model_id || params || prompt) produces the same hash for identical inputs, and the same model produces the same output for the same input (given deterministic sampling).
Cache Proof Verification:
CacheProof = {
requestHash: SHA-256(model_id || params || prompt),
responseHash: SHA-256(response),
proofHash: SHA-256(requestHash || responseHash || metadata),
computedAt: ISO-8601 timestamp,
signature: Ed25519.sign(proofHash, node_private_key),
ttl: 24 hours
}
Verification:
1. Recompute proofHash from (requestHash, responseHash, metadata)
2. Verify Ed25519 signature against node's public key
3. Check TTL has not expired
4. Accept response as authenticated
Popularity Amplification:
When a node fetches a cached response, it becomes a provider for that response. This creates a positive feedback loop where popular responses replicate naturally across the network—the most requested content becomes the most available, following the same principle as BitTorrent's piece selection.
Mechanism 3: KV Prefix Cache (Erasure Coding + KZG)
The KV prefix cache is the most technically sophisticated caching layer, combining information theory (erasure codes) with cryptographic commitments:
Reed-Solomon Erasure Coding:
ERASURE CODING FOR KV STATE
════════════════════════════
Original KV state: 150 MB (Qwen 7B, 2K tokens)
Reed-Solomon(k=32, n=64):
├── 32 data chunks (each ~4.7 MB)
├── 32 parity chunks (each ~4.7 MB)
└── Any 32 of 64 chunks → reconstruct full state
Failure tolerance: up to 50% of chunks can be lost
Distribution: chunks spread across 64 different peers
Recovery: even if 32 peers go offline, state is recoverable
Total redundancy: 2x (300 MB stored across network)
Per-node storage: ~4.7 MB per chunk
KZG Polynomial Commitments:
KZG COMMITMENT SCHEME
══════════════════════
Purpose: verify any chunk of a 600 MB KV cache
with only a 48-byte commitment
1. Encode KV state as polynomial P(x) over finite field
2. Compute commitment C = [P(τ)]₁ (48 bytes, elliptic curve point)
3. Broadcast C via GossipSub
Verification (per chunk):
1. Node requests chunk i from peer
2. Peer provides chunk data + KZG opening proof πᵢ
3. Node verifies: e(C - [vᵢ]₁, [1]₂) = e(πᵢ, [τ - i]₂)
(bilinear pairing check, ~1ms)
4. If valid: chunk is authentic → accept
Properties:
├── 48 bytes to verify any chunk of arbitrarily large state
├── Computational hiding: cannot extract state from commitment
├── Binding: cannot produce valid proof for incorrect chunk
└── Same scheme used in Ethereum danksharding (EIP-4844)
DAP (Data Availability Probing):
DAP PROBING PROTOCOL
════════════════════
Goal: verify that peers actually store the chunks they claim
Method: probes are indistinguishable from real requests
1. Prober sends request for random chunk index
2. Peer must respond with chunk + KZG proof
3. Prober verifies proof
Key insight: because probes look identical to real
requests, a node cannot selectively respond to probes
while withholding from real users.
Failure consequence: reputation penalty
└── Reduces future point earnings
└── Eventually removed from provider lists
Mechanism 4: Fraud Proof Interactive Verification
The fraud proof system implements a bisection-based challenge protocol, analogous to optimistic rollup fraud proofs in blockchain systems:
FRAUD PROOF BISECTION PROTOCOL
═══════════════════════════════
Neural network with 10M neurons (ordered):
Round 1: Submit state roots for each 1M neurons
┌───┬───┬───┬───┬───┬───┬───┬───┬───┬───┐
│1M │2M │3M │4M │5M │6M │7M │8M │9M │10M│
└───┴───┴───┴───┴─×─┴───┴───┴───┴───┴───┘
↑ Challenger: "root #5 is wrong"
Round 2: Subdivide incorrect million into 100K groups
┌────┬────┬────┬────┬────┬────┬────┬────┬────┬────┐
│4.1M│4.2M│4.3M│4.4M│4.5M│4.6M│4.7M│4.8M│4.9M│5.0M│
└────┴────┴────┴──×─┴────┴────┴────┴────┴────┴────┘
↑ Challenger: "root #4 is wrong"
... (continue bisecting) ...
Round 7: Single neuron identified
┌────────────────────────────────────────────┐
│ neuron_4,387,219 │
│ value = activation(Σ weight_j × input_j) │
│ │
│ On-chain verification: │
│ 1. Retrieve weights from Merkle tree │
│ 2. Compute sum + activation │
│ 3. Compare with claimed value │
│ 4. If mismatch → slash challenged node │
└────────────────────────────────────────────┘
Complexity: O(log₁₀(10M)) = 7 rounds
On-chain computation: one neuron evaluation
Mechanism 5: Threshold Signatures and Byzantine Fault Tolerance
THRESHOLD SIGNATURE SCHEME
══════════════════════════
Setup:
├── n nodes, unified public key PK
├── Each node i has unique private key skᵢ
└── Threshold k: any k nodes can produce valid signature
Signing:
1. Node i computes partial signature: ρᵢ ← tsign(skᵢ, m)
2. Collect k partial signatures: {ρᵢ | i ∈ I, |I| = k}
3. Combine: σ ← tcombine(m, {ρᵢ})
4. Verify: tverify(PK, m, σ) → true
Byzantine Fault Tolerance:
├── Up to f = |F| Byzantine nodes
├── k redundant inferences of same query
├── Same result across k nodes → correct with prob 1 - (f/n)^k
├── Different results → fraud proof triggered
└── Honest nodes compensated, dishonest slashed
EigenLayer Integration:
└── Ethereum validators provide cryptoeconomic security
Operators stake ETH → can be slashed for dishonesty
Bootstraps trust without Hyperspace-native token
Mechanism 6: Emergent Cross-Agent Learning
The overnight experiment demonstrated emergent knowledge transfer:
EMERGENT LEARNING PROPAGATION
══════════════════════════════
Timeline of Kaiming Initialization Discovery:
T=0h: Agent A-7f3 begins ML experiments
Tries random init, Xavier init → baseline results
T=2h: Agent A-7f3 tries Kaiming initialization
val_loss drops from 0.961 to 0.942 (1.98% improvement)
T=2h: A-7f3 broadcasts via GossipSub:
"Kaiming init: val_loss 0.961 → 0.942"
T=2.5h: CRDT leaderboard updates across all nodes
A-7f3 moves to rank #1 in ML domain
T=3h: Agent B-2e1 sees Kaiming init in leaderboard
Adopts technique, combines with RMSNorm
T=4h: 8 more agents adopt Kaiming init
T=6h: 15 agents running Kaiming variants
T=8h: 23 of 35 agents have adopted Kaiming init
Multiple variants being explored simultaneously
T=12h: 333 total experiments completed
Knowledge has saturated the network
This is analogous to how ideas spread through a
research community, but compressed from months
(conference cycles) to hours (gossip cycles).
12 Programming Language
Infrastructure Layer
The Hyperspace node infrastructure spans multiple languages:
| Component | Language | Rationale |
|---|---|---|
| aios-cli | Shell (Bash) | Installer scripts, platform detection |
| HyperspaceZK | TypeScript | Zero-knowledge framework, browser integration |
| zkwasm-sdk | TypeScript | WebAssembly ZK SDK |
| Node client | Likely Rust/Go | Performance-critical P2P networking (inferred from libp2p usage) |
| Browser client | TypeScript/JavaScript | WebGPU integration, WebRTC transport |
Research Layer (AGI)
The AGI repo itself functions as a living research artifact rather than a traditional codebase. Agent code runs on the Hyperspace node client, using the OpenAI-compatible API for LLM interactions.
Key Libraries and Dependencies
| Library | Purpose | Used In |
|---|---|---|
| libp2p | P2P networking (DHT, GossipSub, transports) | All node types |
| Loro | CRDT state management | Leaderboard convergence |
| WebGPU/WebLLM | Browser-based GPU inference | Browser nodes |
| Reed-Solomon | Erasure coding for KV cache | Cache distribution |
| Ed25519 | Digital signatures | Cache proofs, identity |
| SHA-256 | Content hashing | Cache keys, Merkle trees |
| KZG | Polynomial commitments | KV chunk verification |
| EigenLayer | Cryptoeconomic security | Validator integration |
Smart Contract Layer
The fraud proof and challenge system involves on-chain components:
ON-CHAIN COMPONENTS
════════════════════
EVM-compatible smart contracts:
├── Fraud proof challenge/response
├── Collateral staking/slashing
├── Threshold signature verification
└── EigenLayer operator registration
Verification:
├── Merkle root validation
├── Single-neuron execution
├── Activation function computation
└── Weight sum verification
Agent-Side Code
Research agents interact with the system through the OpenAI-compatible API, meaning agent logic can be written in any language with an HTTP client. The canonical pattern uses Python:
from openai import OpenAI
client = OpenAI(
base_url="http://localhost:8080/v1",
api_key="hyperspace"
)
def research_loop(domain: str, leaderboard: dict):
"""Continuous research cycle."""
while True:
# 1. Hypothesize (informed by leaderboard + gossip)
hypothesis = client.chat.completions.create(
model="qwen-7b",
messages=[{
"role": "system",
"content": f"Domain: {domain}. Current SOTA: {leaderboard}"
}, {
"role": "user",
"content": "Generate a novel hypothesis and experiment code"
}]
)
# 2. Execute experiment
result = run_experiment(hypothesis.choices[0].message.content)
# 3. Share via gossip + CRDT + Git
broadcast_result(result)
13 Memory Management
Distributed State Architecture
Hyperspace manages state across four distinct layers, each with different consistency, durability, and latency characteristics:
STATE MANAGEMENT HIERARCHY
══════════════════════════
┌─────────────────────────────────────────────────┐
│ Layer 4: PERMANENT ARCHIVE (Git) │
│ ├── Per-agent branches in hyperspaceai/agi │
│ ├── Hourly snapshots (snapshots/latest.json) │
│ ├── Immutable once committed │
│ ├── Human-readable (JSON, Markdown) │
│ └── Latency: ~5 minutes │
├─────────────────────────────────────────────────┤
│ Layer 3: CONVERGENT STATE (Loro CRDTs) │
│ ├── Per-domain leaderboards │
│ ├── Eventually consistent (no coordination) │
│ ├── Delta-only sync between peers │
│ ├── Full snapshot on new node join │
│ └── Latency: ~2 minutes │
├─────────────────────────────────────────────────┤
│ Layer 2: EPHEMERAL CACHE (DHT + Local) │
│ ├── Response cache: SQLite + DHT providers │
│ ├── KV prefix cache: erasure-coded chunks │
│ ├── TTL-based expiration (24h for responses) │
│ ├── Popularity amplification (fetcher → provider)│
│ └── Latency: 1ms (local) to 200ms (DHT) │
├─────────────────────────────────────────────────┤
│ Layer 1: REAL-TIME GOSSIP (GossipSub) │
│ ├── Fire-and-forget messages │
│ ├── Best-effort delivery │
│ ├── Not persisted (consumed or dropped) │
│ ├── Topic-based subscription │
│ └── Latency: ~1 second │
└─────────────────────────────────────────────────┘
Node-Local State
Each node maintains several local data stores:
| Store | Technology | Contents | Persistence |
|---|---|---|---|
| Response Cache | SQLite | SHA-256 indexed responses | Persistent, TTL-evicted |
| KV Prefix State | In-VRAM / on-disk | Attention state for active models | Ephemeral (VRAM) / cached (disk) |
| CRDT Replica | Loro in-memory | Full leaderboard state per domain | Reconstructed from peers on restart |
| Peer Table | k-bucket lists | Known peers by XOR distance | Persistent, refreshed |
| Identity | Keystore | Ed25519 keypair + NodeAddress | Persistent |
| Work Receipts | Local log | Task completion proofs | Persistent until submitted |
Memory Capability: Distributed Vector Store
Nodes with the Memory capability (+5% points) operate a distributed vector store with replication:
DISTRIBUTED VECTOR STORE
═════════════════════════
Agent queries:
"Find experiments similar to 'Kaiming init for transformers'"
│
▼
┌──────────────────────┐
│ Local Embedding │ MiniLM-L6-v2 (CPU)
│ query → vector │ 384-dimensional
└──────────┬───────────┘
│
▼
┌──────────────────────┐
│ Distributed Search │
│ ├── Local index │ Nearest neighbors
│ ├── DHT routing │ Find memory nodes
│ └── Aggregate │ Merge results
└──────────────────────┘
Properties:
├── Replicated across memory-capable nodes
├── Semantic search across all agent experiments
├── Cross-domain knowledge retrieval
└── Used by research loop for hypothesis generation
Cache Hit Rate Dynamics
The cache system exhibits emergent efficiency as network size grows:
CACHE EFFICIENCY vs. NETWORK SIZE
═════════════════════════════════
Hit Rate
100% ┤
│ ╭──────── theoretical max
90% ┤ ╭───╯
│ ╭───╯
80% ┤ ╭───╯
│ ╭───╯
70% ┤ ╭───╯
│ ╭───╯ Zipf's law: popular queries
60% ┤ ╭───╯ dominate, cache fills itself
│╭─╯ with what matters most
50% ┤╯
│
40% ┤
│
30% ┤
└──────┬──────┬──────┬──────┬──────
10K 100K 1M 10M 100M nodes
Key insight: sub-linear cost scaling
├── Linear systems: 2x users = 2x cost
└── Hyperspace: 2x users = higher hit rate = lower cost/user
Garbage Collection and TTL Management
- Response cache: 24-hour TTL on Ed25519-signed cache proofs; entries evicted after TTL or LRU when local storage fills
- KV prefix cache: ephemeral by nature; chunks evicted when VRAM is needed for active inference; on-disk fallback for high-value prefixes
- CRDT state: grows monotonically (append-only leaderboards); compaction via snapshot + truncation of superseded entries
- Gossip messages: consumed on receipt, not persisted; mesh topology ensures delivery without storage overhead
14 Continued Learning
Cross-Agent Knowledge Transfer
Hyperspace AGI's most distinctive continued learning mechanism is its gossip-mediated knowledge transfer. Unlike centralized systems where a knowledge base must be explicitly maintained and queried, Hyperspace leverages the P2P gossip protocol for organic knowledge propagation:
KNOWLEDGE PROPAGATION MODEL
════════════════════════════
Traditional (Centralized):
Agent → Central KB → Agent
├── Single point of failure
├── Bottleneck at scale
└── Curated (editorial overhead)
Hyperspace (Decentralized):
Agent A ←gossip→ Agent B ←gossip→ Agent C
├── No single point of failure
├── Scales with network size
├── Self-curating (leaderboard ranks)
└── Emergent (no editorial overhead)
Multi-Domain Cross-Pollination
Because agents operate across 5 domains with shared gossip infrastructure, discoveries in one domain can inform hypotheses in another:
CROSS-DOMAIN KNOWLEDGE FLOW
════════════════════════════
ML Domain Search Domain
┌─────────────────┐ ┌─────────────────┐
│ "RMSNorm helps │ │ "BM25 + neural │
│ transformer │ ─────→ │ reranker with │
│ training" │ │ normalization │
└─────────────────┘ │ improvements" │
└─────────────────┘
Skills Domain Financial Domain
┌─────────────────┐ ┌─────────────────┐
│ "Improved WASM │ │ "Better data │
│ data extraction" │ ─────→ │ preprocessing │
│ │ │ for backtesting" │
└─────────────────┘ └─────────────────┘
The Causes meta-domain explicitly facilitates
cross-pollination by running sub-causes across
all other domains.
Peer Review and Quality Filtering
The 6-stage research loop includes a peer review mechanism that serves as a quality filter for continued learning:
- Accumulation — Agent runs N experiments, accumulates results
- Synthesis — Agent writes a research paper summarizing findings
- Peer Review — Other agents read, critique, and score papers (1-10 scale)
- Threshold — Papers scoring 8+ are flagged as "breakthroughs"
- Feedback — Breakthrough papers feed back into hypothesis generation
- Evolution — Future agents benefit from curated, high-quality findings
This creates a natural quality gradient: raw experimental results propagate via gossip (fast, unfiltered), while synthesized knowledge propagates via peer review (slower, quality-filtered).
Leaderboard-Driven Curriculum
The CRDT leaderboard implicitly creates a curriculum for new agents:
NEW AGENT ONBOARDING
════════════════════
1. Agent joins network
2. Receives full CRDT leaderboard snapshot (zero cold start)
3. Observes:
├── Current SOTA techniques (top of leaderboard)
├── Active research directions (recent entries)
├── Tried-and-failed approaches (low-ranking entries)
└── Promising unexplored regions (gaps in leaderboard)
4. Agent formulates initial hypothesis informed by
the entire history of network research
5. Begins contributing immediately (no warmup period)
DiLoCo Continuous Model Improvement
When multiple agents collaboratively train models via DiLoCo, the resulting models improve continuously:
- Each agent trains locally for H steps, then shares compressed weight deltas
- Deltas are averaged across participating agents
- The averaged model captures diverse training strategies
- Automatic fallback to solo training if no peers are available
- FP16 gradient all-reduce maintains training quality
Network-Level Learning Curves
At the network level, Hyperspace AGI exhibits a learning dynamic where:
- Early phase: Agents independently explore the search space, discovering low-hanging fruit
- Gossip phase: Successful techniques propagate, agents converge on productive directions
- Refinement phase: With baseline techniques established, agents explore variations and combinations
- Plateau: Diminishing returns as obvious improvements are exhausted
- Breakthrough: Occasional paradigm shifts when an agent discovers a fundamentally new approach
This mirrors the dynamics of human research communities but operates on a compressed timescale (hours vs. months).
Comparison with Other Continued Learning Approaches
| Approach | Hyperspace AGI | FunSearch (DeepMind) | OpenELM | Autoresearch |
|---|---|---|---|---|
| Learning unit | Agent population | Program population | Map population | Single agent |
| Transfer mechanism | P2P gossip + CRDTs | Prompt sampling | Curiosity-driven | None |
| Cross-domain | Yes (5 domains) | No (single problem) | Yes (evolved) | No |
| Persistence | Git + CRDT | Database | Archive | Local logs |
| Quality filter | Peer review (8+ score) | Score-based sampling | Fitness function | Agent judgment |
| Cold start | Zero (CRDT snapshot) | Seed programs | Seed population | From scratch |
| Scale | Thousands of agents | Hundreds of workers | Dozens | One |
15 Applications
Current Applications (Production)
1. Machine Learning Architecture Search
Agents autonomously explore neural network architectures for language modeling on astrophysics papers, optimizing val_loss. This directly extends Karpathy's autoresearch to distributed multi-agent execution.
2. Search Engine Ranking Evolution
Agents evolve BM25 + neural reranker pipelines, optimizing NDCG@10 for web search quality. This combines traditional IR techniques with learned components.
3. Financial Strategy Discovery
Agents backtest S&P 500 monthly-rebalance strategies, optimizing Sharpe ratio. Autonomous exploration of portfolio construction techniques.
4. WASM Skill Forging
Agents develop WebAssembly skills for web scraping, data extraction, and parsing—evaluated by test_pass_rate. This is a self-improvement loop where agents build tools for other agents.
5. Causes Research (Meta-Domain)
A meta-domain that encompasses sub-causes: search ranking improvement, literature analysis, skill forge enhancement, infrastructure development, and data curation.
Infrastructure Applications
6. Decentralized AI Inference
The 2M+ node network serves as a decentralized alternative to centralized AI providers. Any tool compatible with the OpenAI API can use Hyperspace as a backend—replacing API calls to OpenAI/Anthropic/Google with calls to localhost:8080/v1.
7. Distributed Inference Caching
The three-layer cache (response + KV prefix + full inference) eliminates redundant computation at network scale. At 1M nodes, an estimated 70-80% of inference requests are served from cache.
8. Privacy-Preserving AI Access
The proxy capability enables agents to access resources through residential IP proxies, and the fully decentralized architecture means no single entity controls or logs inference requests.
Potential Future Applications
9. Distributed Model Training at Scale
If DiLoCo collaborative training matures, the 2M+ node network could become a distributed training platform for open-weight models—competing with centralized training clusters.
10. Autonomous Scientific Discovery
The multi-domain research framework is domain-agnostic. New research domains can be added by defining a metric and evaluation protocol. Potential domains include: - Drug discovery (binding affinity optimization) - Materials science (property prediction) - Theorem proving (proof success rate) - Code generation (benchmark pass rate)
11. Decentralized AI Research Community
The peer review mechanism (agents scoring papers 8+) could evolve into a decentralized research community where agents produce, review, and build upon each other's work—an autonomous version of the academic publishing process.
12. Federated Knowledge Base
The combination of distributed vector store, CRDT state, and gossip-based knowledge propagation could evolve into a federated knowledge base that grows organically from network research activity.
Significance for the AI Discovery Ecosystem
Hyperspace represents a qualitatively different approach to AI-driven discovery:
POSITIONING IN THE DISCOVERY LANDSCAPE
═══════════════════════════════════════
Single Agent Multi-Agent
Low Scale High Scale
│ │
▼ ▼
┌──────────┐ ┌──────────┐ ┌──────────┐ ┌──────────────┐
│Karpathy │ │FunSearch │ │OpenELM │ │ HYPERSPACE │
│autoresrch│ │(DeepMind)│ │ │ │ AGI │
│ │ │ │ │ │ │ │
│1 agent │ │100s of │ │Diverse │ │ 1000s of │
│1 GPU │ │workers │ │pop'ns │ │ P2P agents │
│1 domain │ │Central │ │MAP-Elites│ │ 5 domains │
│No sharing│ │database │ │Archive │ │ Gossip+CRDT │
│ │ │1 problem │ │ │ │ Git archive │
└──────────┘ └──────────┘ └──────────┘ │ Incentivized │
│ 2M+ nodes │
Centralized ───────────────────────── │ Decentralized│
└──────────────┘
Key differentiation:
├── SCALE: 2M+ nodes vs. single GPU or controlled cluster
├── COORDINATION: gossip + CRDTs vs. central DB/none
├── DOMAINS: 5 simultaneous vs. single problem
├── INCENTIVES: economic alignment vs. none
├── ACCESS: browser tab vs. H100 cluster
└── PERSISTENCE: Git + CRDT vs. local logs
Open Questions for the Field
-
Quality vs. Quantity — Does distributed autoresearch produce genuine insights, or just more parameter sweeps? The Kaiming initialization discovery is encouraging but anecdotal.
-
Incentive Alignment — Does the points economy attract researchers (who optimize for discoveries) or node farmers (who optimize for points)? The +12% research weight is a design-level signal, but real-world behavior may diverge.
-
Emergent Intelligence — Does the gossip-mediated knowledge transfer produce genuinely emergent intelligence (insights no single agent would find alone), or is it merely parallel search with communication?
-
Scaling Laws — How does research quality scale with network size? Linear? Sub-linear? Is there a critical mass beyond which the network becomes self-sustaining?
-
Domain Transfer — Can techniques discovered in one domain (e.g., ML architecture insights) genuinely transfer to another (e.g., financial strategy design), or are the domains too different for cross-pollination?
-
Verification at Scale — As the network grows, can the fraud proof and verification systems maintain integrity, or do adversarial strategies emerge that exploit the decentralized trust model?
These questions position Hyperspace AGI as both a production system and a research experiment—a living laboratory for studying distributed AI-driven discovery at scale.
Analysis based on: Hyperspace whitepaper (hyperspace.computer/bittorrent-for-ai.pdf), GitHub repositories (hyperspaceai/agi, hyperspaceai/hyperspace-node, hyperspaceai/aios-cli), distributed cache specification (cache.hyper.space), Ry Walker Research analysis (rywalker.com/research/hyperspace-agi), Khan et al. 2025 "A Gossip-Enhanced Communication Substrate for Agentic AI" (arXiv:2512.03285), and network dashboard data.