← Back to Index

Hyperspace

Fully decentralized peer-to-peer agent network where thousands of autonomous AI agents collaboratively run experiments, share findings via gossip protocols, and archive breakthroughs—extending single-agent autoresearch into a massively distributed multi-agent discovery system. Organization: HyperspaceAI Published: March 8, 2026 (AGI repo); Protocol whitepaper ~2023 Type: repo + whitepaper + production network Report Type: PhD-Level Technical Analysis Report Date: April 2026

Table of Contents


1 Full Title and Attribution

Full Title: Hyperspace: A Peer-to-Peer Artificial Intelligence Network

Extended Title (AGI layer): Hyperspace AGI — The First Distributed AGI System: Autonomous AI Agents Collaboratively Train Models, Share Experiments via P2P Gossip, and Push Breakthroughs

Whitepaper: hyperspace.computer/bittorrent-for-ai.pdf

AGI Repository: github.com/hyperspaceai/agi (1,238 stars, MIT license)

Node Repository: github.com/hyperspaceai/hyperspace-node (258 stars)

CLI Repository: github.com/hyperspaceai/aios-cli (80 stars, 2.7M+ downloads)

GitHub Organization: github.com/hyperspaceai (13 public repos, founded June 2015)

Network Dashboard: network.hyper.space

Agent Portal: agents.hyper.space

Distributed Cache Spec: cache.hyper.space

"Intelligence compounds when thousands of agents share findings in real-time."

Hyperspace is a two-layer system. The infrastructure layer is a production-grade decentralized P2P AI inference network with 2M+ active nodes and 3.6M+ client downloads, built on libp2p/IPFS protocols. The research layer (Hyperspace AGI) sits on top of this infrastructure, implementing a distributed autoresearch system where thousands of autonomous agents collaboratively run experiments, gossip findings, maintain convergent leaderboards via CRDTs, and archive results to GitHub. The AGI repo was created on March 8, 2026—two days after Karpathy's seminal autoresearch release—scaling the pattern from single-agent/single-GPU to a massively distributed multi-agent system.

The system is positioned explicitly as "BitTorrent for AI inference" and represents the most ambitious attempt at distributed AI-driven discovery in the open-source ecosystem, combining distributed systems theory (DHTs, gossip protocols, CRDTs, erasure coding), cryptographic verification (fraud proofs, threshold signatures, KZG commitments), and agentic AI research loops into a unified production platform.


2 Authors and Team

Core Team

Role Name Background
CEO & Founder Ohad Levi Previously WAY2VAT, ClickSoftware, HP, Intel
CTO & Co-Founder Max Nigri Systems architecture, distributed systems

Organization Profile

Attribute Value
Organization HyperspaceAI
Location United States of America
GitHub Org Created June 1, 2015
GitHub Followers 596
Public Repos 13
Network Scale 2M+ active nodes, 3.6M+ client downloads
Contact gutenberg@hyperspace.us

Intellectual Lineage

Hyperspace draws from multiple intellectual traditions simultaneously:

  1. P2P Systems — BitTorrent's tit-for-tat economics, IPFS's content-addressed storage, Kademlia DHT routing
  2. Blockchain/Crypto — EigenLayer restaking for cryptoeconomic security, fraud proofs from optimistic rollups, threshold signatures
  3. Distributed Systems Theory — CRDTs (conflict-free replicated data types), gossip protocols, eventual consistency
  4. AI Research Automation — Karpathy's autoresearch pattern (March 6, 2026), extended to multi-agent distributed setting
  5. Information Retrieval — BM25 + neural reranking for search domain experiments

The gossip-for-agents paradigm is formalized in a companion academic paper:

Khan, N.I., Habiba, M., & Khan, R. (2025). "A Gossip-Enhanced Communication Substrate for Agentic AI: Toward Decentralized Coordination in Large-Scale Multi-Agent Systems." arXiv:2512.03285

This paper provides theoretical foundations for gossip protocols as coordination substrates in agentic systems, arguing that gossip is "essential for future agentic ecosystems that must remain robust, adaptive, and self-organizing as their scale and autonomy increase."

Third-Party Analysis

The system has been analyzed by Ry Walker Research (rywalker.com), which describes it as "the most ambitious entry in the autoresearch category" and "the most sophisticated multi-agent research coordination system in the open-source ecosystem."


3 Core Contribution

The Problem: Single-Agent Autoresearch Doesn't Scale

Karpathy's autoresearch (March 6, 2026) proved that a single LLM agent on a single GPU can autonomously generate hypotheses, run ML training experiments, and improve model architectures. The system achieved impressive results—reducing val_loss from 0.9979 to 0.9697 in 125 experiments—but is fundamentally constrained:

SINGLE-AGENT AUTORESEARCH (Karpathy)
════════════════════════════════════════════════════════

  ┌─────────────────────────┐
  │   Single Agent          │
  │   ┌───────────────────┐ │
  │   │ Hypothesize       │ │
  │   │ Experiment        │ │
  │   │ Analyze           │ │
  │   │ Repeat            │ │
  │   └───────────────────┘ │
  │                         │
  │   Single GPU (H100)     │
  │   Single Domain (LLM)   │
  │   No Communication      │
  │   No Knowledge Sharing  │
  └─────────────────────────┘

  Constraints:
  ├── Compute: bounded by one GPU
  ├── Ideas: bounded by one agent's exploration
  ├── Domains: LLM training only
  └── Learning: no cross-pollination of discoveries

The Solution: Distributed Multi-Agent Research via P2P Gossip

Hyperspace AGI extends autoresearch from a single researcher to an entire research community—thousands of agents collaborating across multiple domains with real-time knowledge sharing:

DISTRIBUTED AUTORESEARCH (Hyperspace AGI)
════════════════════════════════════════════════════════

  Agent A ←──gossip──→ Agent B ←──gossip──→ Agent C
    │                    │                    │
    ▼                    ▼                    ▼
  ┌──────┐           ┌──────┐           ┌──────┐
  │ ML   │           │Search│           │ Fin  │
  │Domain│           │Engine│           │ Anal │
  └──┬───┘           └──┬───┘           └──┬───┘
     │                   │                   │
     └───────┬───────────┴───────────┬───────┘
             │                       │
             ▼                       ▼
  ┌──────────────────┐   ┌──────────────────┐
  │  CRDT Leaderboard │   │  GitHub Archive  │
  │  (convergent      │   │  (permanent      │
  │   state, ~2min)   │   │   record, ~5min) │
  └──────────────────┘   └──────────────────┘

  Advantages:
  ├── Compute: distributed across 2M+ nodes
  ├── Ideas: cross-pollination via gossip (~1s)
  ├── Domains: 5 simultaneous research tracks
  ├── State: convergent via CRDTs (zero cold start)
  └── Archive: permanent via Git (reproducible)

Three Fundamental Contributions

Contribution 1: Three-Layer Coordination Stack

The primary architectural innovation is a tiered coordination architecture with latency-appropriate consistency guarantees:

Layer Protocol Latency Consistency Purpose
1. Gossip libp2p GossipSub ~1 second Best-effort broadcast Real-time inspiration: "Agent X found Kaiming init helps"
2. CRDT Loro CRDTs ~2 minutes Eventual convergence Authoritative state: per-domain leaderboards with rankings
3. Archive Git + GitHub ~5 minutes Durable, immutable Permanent record: per-agent branches, hourly snapshots

This layering solves a real distributed systems problem: gossip provides speed for fast iteration cycles, CRDTs provide consistency without coordination overhead, and Git provides durability and human-readable reproducibility.

Contribution 2: Research-as-Portfolio

Rather than treating autoresearch as a single-domain problem, Hyperspace AGI frames it as a portfolio optimization problem across five simultaneous domains. Agents can specialize (deep expertise in ML training) or generalize (broad exploration across domains), and findings in one domain can cross-pollinate others.

Contribution 3: Incentive-Aligned Distributed Compute for Research

The points economy aligns individual incentives with collective research progress. The research capability carries the highest point weight (+12%), explicitly signaling the network's priority on autoresearch over mere inference serving.

Comparison with Karpathy's autoresearch

Dimension Karpathy autoresearch Hyperspace AGI
Agents Single Thousands
Compute Single GPU 2M+ distributed nodes
Domains LLM training only 5 simultaneous domains
Coordination None (solo) GossipSub + CRDT + GitHub
Knowledge sharing None Real-time gossip (~1s)
State convergence N/A Loro CRDTs (~2min)
Reproducibility Local logs Per-agent Git branches
Barrier to entry H100 + coding agent Browser tab
Stars (GitHub) ~30,000 ~1,238
Complexity 3 files, 630 lines Full P2P protocol stack
Collaborative training No DiLoCo (compressed weight deltas)

4 Supported Solutions

Research Domains

Hyperspace AGI operates across five simultaneous research domains, each with its own optimization metric, CRDT leaderboard, and agent population:

Domain Metric Direction Agent Activity
Machine Learning val_loss ↓ lower is better Train language models on astrophysics papers (Karpathy-style). Architecture search, hyperparameter optimization, initialization strategies
Search Engine NDCG@10 ↑ higher is better Evolve BM25 + neural reranker pipelines for web search quality
Financial Analysis Sharpe ratio ↑ higher is better Backtest S&P 500 monthly-rebalance strategies, optimize risk-adjusted returns
Skills & Tools test_pass_rate ↑ higher is better Forge WASM skills for web scraping, parsing, data extraction
Causes Per-cause metric varies Meta-domain: search ranking, literature analysis, skill forge, infrastructure, data curation

The Research Loop

Each agent runs a continuous six-stage cycle inspired by Karpathy's autoresearch but extended with collaborative elements:

THE HYPERSPACE RESEARCH LOOP
═══════════════════════════════════════════════

  ┌─────────────────────────────────────────┐
  │                                         │
  │  1. HYPOTHESIZE                         │
  │     Generate research ideas             │
  │     "What if RMSNorm > LayerNorm?"      │
  │     Informed by gossip from peers       │
  │                                         │
  │  2. EXPERIMENT                          │
  │     Run on available hardware           │
  │     (browser tab ──── to ──── H100)     │
  │                                         │
  │  3. SHARE                               │
  │     Broadcast via GossipSub (~1s)       │
  │     Update CRDT leaderboard (~2min)     │
  │                                         │
  │  4. SYNTHESIZE                          │
  │     Accumulate N experiments            │
  │     Generate research paper             │
  │                                         │
  │  5. PEER REVIEW                         │
  │     Other agents read + critique        │
  │     Score 1-10                          │
  │                                         │
  │  6. EVOLVE                              │
  │     Papers scoring 8+ → "breakthrough"  │
  │     Feed back into Stage 1              │
  │                                         │
  └─────────────────────────────────────────┘
                    │
                    ▼
              [Loop forever]

Collaborative Training (DiLoCo)

Multiple agents can collaboratively train the same model using the Distributed Low-Communication (DiLoCo) protocol:

DiLoCo COLLABORATIVE TRAINING
═══════════════════════════════

  Agent A              Agent B              Agent C
  ┌──────────┐        ┌──────────┐        ┌──────────┐
  │ Train H  │        │ Train H  │        │ Train H  │
  │ steps    │        │ steps    │        │ steps    │
  │ locally  │        │ locally  │        │ locally  │
  └────┬─────┘        └────┬─────┘        └────┬─────┘
       │                   │                   │
       ▼                   ▼                   ▼
  Compress            Compress            Compress
  weight              weight              weight
  deltas              deltas              deltas
       │                   │                   │
       └───────────────────┼───────────────────┘
                           │
                    ┌──────▼──────┐
                    │   Average   │
                    │   deltas    │
                    │   via P2P   │
                    └──────┬──────┘
                           │
                    Apply to model
                           │
                    [Repeat cycle]

  Properties:
  ├── 90-95% compute utilization
  ├── Works across continents
  ├── Automatic fallback to solo training
  └── FP16 gradient all-reduce (no perf loss)

Deployment Modes

Client Hardware Performance Use Case
Browser (agents.hyper.space) WebGPU 10-20 tok/s Zero-install participation, smaller models
CLI Native GPU 40-80 tok/s Full performance, all model sizes
Tray App Desktop GPU 40-80 tok/s Background daemon with auto-updates
Headless Server GPU 40-80+ tok/s Cloud VMs, dedicated GPU servers

5 LLM Integration

OpenAI-Compatible Local API

Every Hyperspace node exposes an OpenAI-compatible API at localhost:8080/v1, enabling seamless integration with any tool in the LLM ecosystem:

┌────────────────────────────────────────────────────┐
│            OpenAI-Compatible API Surface           │
│                                                    │
│  Base URL: http://localhost:8080/v1                 │
│                                                    │
│  Endpoints:                                        │
│  ├── POST /chat/completions  (text generation)     │
│  ├── GET  /models            (available models)    │
│  └── POST /embeddings        (vector embeddings)   │
│                                                    │
│  Compatible with:                                  │
│  ├── OpenAI Python SDK (change base_url)           │
│  ├── OpenAI TypeScript SDK                         │
│  ├── LangChain                                     │
│  ├── LlamaIndex                                    │
│  ├── Any OpenAI-compatible tool                    │
│  └── AutoResearch pipelines                        │
└────────────────────────────────────────────────────┘

Model Serving Architecture

The inference pipeline operates in a fully decentralized manner:

CLIENT REQUEST
      │
      ▼
┌─────────────────────┐
│  Local API Gateway  │
│  localhost:8080/v1   │
└─────────┬───────────┘
          │
          ▼
┌─────────────────────────────────────────────┐
│           DISTRIBUTED CACHE CHECK           │
│                                             │
│  Layer 1: Response Cache                    │
│  SHA-256(prompt + model + params) → SQLite  │
│  Hit? → Return verified cached response     │
│                                             │
│  Layer 2: KV Prefix Cache                   │
│  Same prefix tokens → skip prefill          │
│  Erasure-coded KV state across peers        │
│                                             │
│  Layer 3: Full Inference                    │
│  Route to peer with best capacity           │
│  Hedged requests for tail latency           │
└─────────────────┬───────────────────────────┘
                  │
                  ▼
┌─────────────────────────────────────────────┐
│           P2P MODEL INFERENCE               │
│                                             │
│  Registry → DHT → PEX → Gossip routing     │
│  Model loaded on local/remote GPU           │
│  Result → store in L1 + L2 caches           │
│  Announce via GossipSub + DHT               │
└─────────────────────────────────────────────┘

Embedding Support

Nodes with the Embedding capability run CPU-based vector embeddings using MiniLM-L6-v2:

  • Used for semantic cache matching (cosine similarity, 0.92 threshold)
  • Enables "What is ML?" to match "Explain machine learning" without re-inference
  • Supports the distributed memory capability for agent knowledge stores

Model Availability

The network supports open-weight models that can be served by consumer and enterprise hardware:

Model Tier Examples Minimum Hardware
Small (0.5-3B) Qwen 0.5B Browser (WebGPU)
Medium (7-14B) Qwen 7B, Gemma-3 8 GB GPU
Large (27-32B) Gemma-3 27B, Qwen 32B 24-48 GB GPU
XL (70B+) Qwen 72B 80 GB GPU

Agent-LLM Interaction Pattern

The research agents use LLMs in a tightly-coupled loop where the LLM is both the idea generator (hypothesis formation) and the executor (code generation for experiments):

AGENT ←→ LLM INTERACTION
═════════════════════════

  ┌──────────────────────────────────────────────┐
  │  Agent Control Loop                          │
  │                                              │
  │  1. Read CRDT leaderboard (current SOTA)     │
  │  2. Read gossip feed (peer discoveries)      │
  │  3. Prompt LLM:                              │
  │     "Given these results, hypothesize a       │
  │      new architecture modification"           │
  │  4. LLM generates hypothesis + code          │
  │  5. Agent executes experiment                │
  │  6. Agent evaluates results                  │
  │  7. If improved: broadcast via GossipSub     │
  │  8. Update CRDT leaderboard entry            │
  │  9. Push to per-agent Git branch             │
  │  10. Goto 1                                  │
  └──────────────────────────────────────────────┘

6 Key Results

Network-Scale Metrics

Metric Value Context
Active nodes 2,000,000+ Largest decentralized AI inference network
Client downloads 3,600,000+ CLI + Tray App + Browser
GitHub stars (AGI) 1,238 Created March 8, 2026
GitHub stars (node) 258 Production infrastructure
Bootstrap nodes 6 US, EU, Asia, South America, Oceania
Research domains 5 Simultaneous multi-domain operation

Overnight Proof-of-Concept (March 2026)

The most documented experimental result involves 35 agents running 333 experiments autonomously overnight:

Metric Value
Agents 35
Total experiments 333
Duration One night (~8-12 hours)
Key discovery Kaiming initialization improves val_loss
Adoption speed 23 peers adopted within hours
Coordination mechanism GossipSub broadcast → CRDT convergence

Emergent behavior observed: When one agent discovered that Kaiming initialization improved training outcomes, 23 out of 35 peers adopted the technique within hours via the gossip protocol. This demonstrates emergent learning without centralized coordination—successful strategies propagate organically through the network.

Snapshot Data (Ongoing)

The AGI repository publishes consolidated network snapshots (snapshots/latest.json) every hour:

  • Raw CRDT leaderboard state from the live network
  • Per-agent branch commits show individual experiment histories
  • Active daily commits from network agents
  • As of March 2026: 67 agents running 1,369+ experiments across active domains

Distributed Cache Performance Claims

Network Size Combined Hit Rate Energy Saved/Year CO₂ Avoided/Year
10K nodes 30-45% 110-165 MWh 47-71 tons
100K nodes 50-70% 7,300-15,300 MWh 3,100-6,600 tons
1M nodes 65-80% 142,000-175,000 MWh 61,000-75,000 tons
10M nodes 75-90% 411,000-493,000 MWh 176,000-211,000 tons

Assumptions: 50 req/node/day, ~3 Wh/inference on consumer GPU, 0.429 kg CO₂/kWh global average.

Competitive Positioning

System Agents Compute Domains Coordination Stars
Karpathy autoresearch Single Single GPU LLM training None ~30,000
autoresearch-at-home Conceptual Planned distributed Planned Conceptual ~188
Hyperspace AGI Thousands 2M+ P2P nodes 5 domains GossipSub + CRDT + Git ~1,238

Strengths Identified by Third-Party Analysis

  1. Network effects moat — 2M+ nodes is unmatched; no other autoresearch system has comparable distributed compute
  2. Multi-domain research — 5 simultaneous research tracks vs. single-domain competitors
  3. Production P2P infrastructure — Real libp2p network with DHT, gossip, NAT traversal (not a prototype)
  4. Zero-install browser option — Lowest barrier to entry in the category
  5. Hourly GitHub snapshots — Full transparency via raw CRDT state

Weaknesses and Open Questions

  1. Research quality unproven — No published breakthrough results beyond the Kaiming initialization anecdote; unclear if distributed agents produce genuine insights or mechanical parameter sweeps
  2. Token/crypto adjacency — Points system and "earn while you compute" framing may alienate pure research audiences
  3. Complexity — 9 capability types, 5 research domains, 3 coordination layers, verification protocols—many moving parts
  4. Agent autonomy unclear — How much genuine research insight vs. brute-force hyperparameter search?

7 Reproducibility

High Reproducibility by Design

Hyperspace AGI is architecturally designed for reproducibility through its three-layer archival system:

Layer 1: Per-Agent Git Branches

Every agent pushes experiment results to its own branch in hyperspaceai/agi. Each commit contains: - Experiment configuration - Model architecture details - Training results (metrics, loss curves) - The specific hypothesis being tested

Layer 2: Hourly Snapshots

A network node publishes consolidated snapshots to snapshots/latest.json every hour. These contain the raw CRDT leaderboard state, providing a time-series record of the network's collective research progress.

Layer 3: Raw CRDT State

The Loro CRDT leaderboard state is itself a reproducible artifact. Because CRDTs are deterministic given the same set of operations, any node that receives all operations will converge to the same state—regardless of message ordering or network partitions.

Installation Reproducibility

# One-line CLI install (auto-detects platform)
curl -fsSL https://download.hyper.space/api/install | bash

# Platform-specific options:
# macOS:  .dmg tray app
# Linux:  .deb package
# Windows: .exe / .msi installer
# Headless: --no-tray flag for servers

# Browser: zero-install at agents.hyper.space

OpenAI API Compatibility

Because the local API is OpenAI-compatible, any existing autoresearch pipeline can be trivially pointed at a Hyperspace node:

from openai import OpenAI

client = OpenAI(
    base_url="http://localhost:8080/v1",
    api_key="hyperspace"
)

response = client.chat.completions.create(
    model="qwen-7b",
    messages=[{"role": "user", "content": "Hypothesize..."}]
)

Limitations on Reproducibility

  1. Non-deterministic gossip ordering — Which agent sees which discovery first is a function of network topology and timing
  2. Hardware heterogeneity — Experiments run on hardware from browser WebGPU to H100s, introducing variance
  3. Agent population dynamics — The set of active agents changes over time as nodes join and leave
  4. Model availability — Not all models are available on all hardware classes (browser nodes are limited to small models)

8 Compute and API Costs

Points Economy

Hyperspace uses a dual-stream incentive system rather than direct monetary compensation:

Stream 1: Presence Points (Passive)

Every 90-second pulse round: - Base: 10 points - Uptime bonus: U(t) = 1 + 0.2 × ln(1 + t/12) - 30-day continuous uptime → 83% bonus - Capability bonuses: per capability enabled

Stream 2: Work Points (Active)

Per task completed: - Formula: tokens × cost_per_token × model_multiplier × uptime_bonus - Research experiments count as work (highest-weighted capability)

Earnings by Setup

Setup Duty Cycle Points/Day Points/Month Equivalent Tier
Browser, 2h/day ~8% ~19 ~460 Casual contributor
Browser, 24h/day 100% ~228 ~5,600 Dedicated browser
Desktop, 8GB GPU 100% ~503 ~12,800 Consumer hardware
Server, 80GB GPU 100% ~1,912 ~44,100 Enterprise hardware

Capability Point Weights

Capability Point Weight Cost Profile
Research +12% GPU-intensive (training experiments)
Inference +10% GPU-intensive (model serving)
Proxy +8% Network bandwidth
Storage +6% Disk I/O
Embedding +5% CPU-intensive
Memory +5% RAM + disk
Orchestration +5% CPU + networking
Validation +4% CPU (proof verification)
Relay +3% Network bandwidth (NAT traversal)

Research (+12%) is the highest-weighted capability, signaling that the network's economic priority is autoresearch.

Cost Comparison: Centralized vs. Distributed

The distributed cache architecture fundamentally changes the cost curve:

COST SCALING COMPARISON
═══════════════════════

  Cost
   │
   │                      ╱ Centralized (Linear)
   │                    ╱   2x users = 2x GPU cost
   │                  ╱     Every query is a cost center
   │                ╱       KV cache siloed per datacenter
   │              ╱
   │            ╱        ╱── Hyperspace (Sub-linear)
   │          ╱       ╱      2x users = higher hit rate
   │        ╱      ╱         Every query enriches cache
   │      ╱    ╱             KV state shared globally
   │    ╱  ╱
   │  ╱╱
   │╱╱
   └──────────────────────────────────── Users

  At 1M nodes: only 6-30% of requests need full inference
  At 10M nodes: 75-90% served from cache

No Direct API Cost

Unlike cloud LLM providers, there are no per-token API costs. The economic model is:

  1. Compute contribution — Users donate GPU cycles
  2. Point accumulation — Points earned proportional to contribution
  3. Network benefit — Access to distributed inference and research network
  4. Cache economics — Marginal cost approaches zero as cache fills

9 Architecture Solution

System Overview: Two-Layer Architecture

Hyperspace is a two-layer system where each layer can be understood independently:

HYPERSPACE SYSTEM ARCHITECTURE
════════════════════════════════════════════════════════════

  ┌──────────────────────────────────────────────────────┐
  │                 RESEARCH LAYER                       │
  │                (Hyperspace AGI)                      │
  │                                                      │
  │  ┌────────┐  ┌────────┐  ┌────────┐  ┌────────┐    │
  │  │Agent 1 │  │Agent 2 │  │Agent 3 │  │Agent N │    │
  │  │ML      │  │Search  │  │Finance │  │Skills  │    │
  │  └───┬────┘  └───┬────┘  └───┬────┘  └───┬────┘    │
  │      │           │           │           │          │
  │      └─────┬─────┴─────┬─────┴─────┬─────┘          │
  │            │           │           │                 │
  │     ┌──────▼─────┐ ┌──▼────────┐ ┌▼──────────┐     │
  │     │ GossipSub  │ │   CRDT    │ │  GitHub    │     │
  │     │ (~1 sec)   │ │ (~2 min)  │ │ (~5 min)   │     │
  │     │ broadcast  │ │ converge  │ │ archive    │     │
  │     └────────────┘ └───────────┘ └────────────┘     │
  │                                                      │
  └──────────────────────┬───────────────────────────────┘
                         │ built on top of
  ┌──────────────────────▼───────────────────────────────┐
  │              INFRASTRUCTURE LAYER                    │
  │             (Hyperspace Network)                     │
  │                                                      │
  │  ┌───────────────────────────────────────────────┐   │
  │  │              libp2p Protocol Stack             │   │
  │  │  ┌─────────┐ ┌──────────┐ ┌────────────────┐  │   │
  │  │  │Kademlia │ │GossipSub │ │NAT Traversal   │  │   │
  │  │  │DHT      │ │Pub/Sub   │ │+ Relay         │  │   │
  │  │  └─────────┘ └──────────┘ └────────────────┘  │   │
  │  └───────────────────────────────────────────────┘   │
  │                                                      │
  │  ┌───────────────────────────────────────────────┐   │
  │  │            Distributed Cache                  │   │
  │  │  L1: Response Cache (SHA-256 + Ed25519)       │   │
  │  │  L2: KV Prefix Cache (Reed-Solomon + KZG)     │   │
  │  │  L3: Full Inference (P2P routing)             │   │
  │  └───────────────────────────────────────────────┘   │
  │                                                      │
  │  ┌───────────────────────────────────────────────┐   │
  │  │           Cryptoeconomic Layer                │   │
  │  │  Pulse rounds (90s) + Work receipts           │   │
  │  │  Fraud proofs + Challenge model               │   │
  │  │  EigenLayer operator integration              │   │
  │  └───────────────────────────────────────────────┘   │
  │                                                      │
  │  2,000,000+ nodes  ·  3,600,000+ downloads          │
  │  6 bootstrap nodes  ·  Global coverage               │
  └──────────────────────────────────────────────────────┘

Network Protocol Stack

The infrastructure layer is built on battle-tested distributed systems primitives:

PROTOCOL STACK
══════════════

  ┌─────────────────────────────────┐
  │    Application (Agent Logic)     │
  ├─────────────────────────────────┤
  │    OpenAI API (/v1/*)            │  ← Standard interface
  ├─────────────────────────────────┤
  │    Distributed Cache (3-layer)   │  ← Response + KV + Routing
  ├─────────────────────────────────┤
  │    CRDT State (Loro)             │  ← Convergent leaderboards
  ├─────────────────────────────────┤
  │    GossipSub Pub/Sub             │  ← Topic-based broadcast
  ├─────────────────────────────────┤
  │    Kademlia DHT                  │  ← Peer/content discovery
  ├─────────────────────────────────┤
  │    S/Kademlia + Crypto Puzzles   │  ← Sybil resistance
  ├─────────────────────────────────┤
  │    libp2p Transport              │  ← TCP, WebSocket, WebRTC
  ├─────────────────────────────────┤
  │    NAT Traversal + Relay         │  ← Browser node support
  └─────────────────────────────────┘

GossipSub Topics and Protocol Identifiers

Type Identifier Purpose
GossipSub hyperspace/cache/announcements Response cache availability
GossipSub hyperspace/kv-prefix/announcements KV prefix availability
Protocol /hyperspace/cache/1.0.0 Response cache fetch
Protocol /hyperspace/kv-cache/1.0.0 KV chunk fetch + prefix routing
DHT Key /cache/<hash> Response cache provider discovery
DHT Key /kv-prefix/<model>/<prefix-hash> KV prefix provider discovery

Identity and Security Model

From the whitepaper, the identity system is designed for Sybil resistance in a fully decentralized environment:

NODE IDENTITY SYSTEM
═══════════════════

  1. KeyPair Generation
     Ed25519 public/private key pair

  2. NodeAddress = Hash(PublicKey)
     Cryptographic hash prevents spoofing

  3. Proof of Work Challenge
     Computationally expensive ID generation
     Prevents Sybil attacks at scale

  4. Message Signing (two tiers):
     ├── Weak: sign(IP, port, timestamp)
     │   For non-critical messages (PING)
     │   Counters replay attacks
     │
     └── Strong: sign(full_message + nonce)
         For critical messages
         Full integrity + Man-in-Middle protection

  5. Join Protocol:
     Node → HCS: join(address, weak_sig)
     HCS → Node: crypto_puzzle(complexity c2)
     Node → HCS: registry(capabilities, models)
     HCS → Node: inference_fortification_challenge
     Node → HCS: verify_inference(result)

10 Component Breakdown

Node Capability Matrix

Each Hyperspace node can run any combination of 9 capabilities, creating a heterogeneous network of specialized and generalist nodes:

# Capability Function Resource Profile Point Weight
1 Inference Serve AI models (GPU-accelerated) GPU (4-80 GB VRAM) +10%
2 Research Run ML training experiments GPU + CPU + Storage +12%
3 Proxy Residential IP proxy for agents Network bandwidth +8%
4 Storage DHT block storage Disk I/O, capacity +6%
5 Embedding CPU vector embeddings (MiniLM-L6-v2) CPU-intensive +5%
6 Memory Distributed vector store with replication RAM + disk +5%
7 Orchestration Multi-step task decomposition + routing CPU + networking +5%
8 Validation Verify proofs in pulse rounds CPU (verification) +4%
9 Relay NAT traversal for browser nodes Network bandwidth +3%

Component 1: Kademlia DHT Layer

The DHT layer handles peer discovery and content routing across the network:

KADEMLIA DHT (Extended)
═══════════════════════

  Properties:
  ├── XOR metric for distance calculation
  ├── O(log n) lookup hops
  ├── k-bucket peer lists by distance
  ├── S/Kademlia extensions:
  │   ├── Cryptographic puzzle for ID generation
  │   ├── Public key infrastructure per node
  │   ├── Signed messages for trustworthy routing
  │   ├── Regular ID refresh (anti long-term Sybil)
  │   └── Mandatory bucket updates (anti Eclipse)
  └── Suzaku enhancements:
      ├── Bidirectional finger table
      ├── [log₂ n] - 1 max lookup hops
      └── CHURN-resilient (lockless liveness)

  Use in Hyperspace:
  ├── /cache/<hash>           → find response cache providers
  ├── /kv-prefix/<m>/<hash>   → find KV prefix providers
  └── General peer discovery  → node-to-node connections

Component 2: GossipSub Messaging

libp2p's GossipSub provides topic-based publish-subscribe messaging:

GOSSIPSUB IN HYPERSPACE
═══════════════════════

  Topics:
  ├── hyperspace/cache/announcements
  │   └── "I have cached response for SHA-256(X)"
  │
  ├── hyperspace/kv-prefix/announcements
  │   └── "I have KV state for model Y, prefix Z"
  │
  └── [Research domain topics]
      └── "Agent A found: Kaiming init → val_loss -3%"

  Properties:
  ├── ~1 second message propagation
  ├── Mesh-based topology (not flood)
  ├── Peer scoring for spam resistance
  └── Subscription-based filtering

Component 3: Loro CRDT Leaderboards

Each research domain maintains a conflict-free replicated data type (CRDT) leaderboard:

CRDT LEADERBOARD ARCHITECTURE
══════════════════════════════

  Domain: Machine Learning
  ┌─────────────────────────────────────────────┐
  │  Rank  │  Agent  │  val_loss  │  Technique  │
  ├────────┼─────────┼────────────┼─────────────┤
  │  1     │  A-7f3  │  0.9412    │  Kaiming+RMS│
  │  2     │  B-2e1  │  0.9523    │  Kaiming    │
  │  3     │  C-9a4  │  0.9601    │  LayerNorm  │
  │  ...   │  ...    │  ...       │  ...        │
  └─────────────────────────────────────────────┘

  Properties (Loro CRDTs):
  ├── Convergent: all nodes see same final state
  ├── Commutative: order of operations irrelevant
  ├── Zero cold start: new nodes read full state
  ├── Delta-only sync: minimal bandwidth
  ├── ~2 minute convergence across network
  └── No coordination required (no consensus)

Component 4: Distributed Response Cache (Layer 1)

RESPONSE CACHE FLOW
════════════════════

  Request → SHA-256(modelId + params + prompt)
                │
                ▼
  ┌──────────────────────┐
  │  1. Local SQLite     │ → Hit? Return (~1ms)
  └──────────┬───────────┘
             │ miss
             ▼
  ┌──────────────────────┐
  │  2. DHT Lookup       │ → findProviders('/cache/' + hash)
  │     for providers    │
  └──────────┬───────────┘
             │ found
             ▼
  ┌──────────────────────┐
  │  3. Stream Fetch     │ → /hyperspace/cache/1.0.0
  │     from peer        │
  └──────────┬───────────┘
             │
             ▼
  ┌──────────────────────┐
  │  4. Verify Proof     │
  │  CacheProof:         │
  │    SHA-256(reqHash    │
  │    || respHash       │
  │    || proofHash      │
  │    || computedAt)    │
  │  Ed25519 signed      │
  │  24h TTL             │
  └──────────┬───────────┘
             │ valid
             ▼
  ┌──────────────────────┐
  │  5. Store + Announce │
  │  Local SQLite cache  │
  │  DHT re-announce     │   ← Popularity amplification
  │  (fetchers become    │
  │   providers)         │
  └──────────────────────┘

  Semantic matching:
  └── Cosine similarity (0.92 threshold)
      "What is ML?" ≈ "Explain machine learning"

Component 5: KV Prefix Cache (Layer 2)

KV PREFIX CACHE
═══════════════

  Why it matters:
  ├── Prefill is the most expensive part of inference
  ├── Same model + same tokens = same KV state (deterministic)
  └── Zero reason to compute it more than once globally

  KV State Sizes:
  ┌──────────────────────────────────────────────┐
  │  Model        │ 512 tok │ 2K tok  │ 8K tok  │
  ├───────────────┼─────────┼─────────┼─────────┤
  │  Qwen 0.5B    │   4 MB  │  15 MB  │  60 MB  │
  │  Qwen 7B      │  38 MB  │ 150 MB  │ 600 MB  │
  │  Gemma-3 27B  │ 150 MB  │ 600 MB  │ 2.4 GB  │
  │  Qwen 32B     │ 175 MB  │ 700 MB  │ 2.8 GB  │
  └──────────────────────────────────────────────┘

  Distribution Protocol:
  1. Node A computes KV state → exports via context.getStateData()
  2. Reed-Solomon erasure coding: k=32, n=64 chunks
  3. KZG polynomial commitment (48 bytes)
  4. Chunks pushed via routing network
  5. Commitment gossiped across P2P mesh
  6. DAP-style availability probing

  Verification:
  ├── KZG commitment:  48 bytes to verify any chunk
  ├── Any 32 of 64 chunks reconstructs full state
  ├── DAP probing:     indistinguishable from real requests
  └── Failure:         reputation penalty

Component 6: Fraud Proof System

From the whitepaper, Hyperspace implements an interactive fraud proof mechanism:

FRAUD PROOF PROTOCOL
════════════════════

  1. Client detects two different replies for same query

  2. Client submits fraud claim on-chain

  3. Challenge process begins (logarithmic rounds):

     Round 1: Challenged node submits 10 state roots
              (neurons 1M at a time)

     Round 2: Challenger identifies first incorrect root
              Submits challenge

     Round 3: Challenged node submits 10 sub-roots
              (100K neurons each)

     ...continues bisecting...

     Round 7: Single neuron identified
              On-chain verification:
              sum(incoming_weights × values) + activation

  Properties:
  ├── O(log n) messages for n execution steps
  ├── Final verification is simple arithmetic on-chain
  ├── Honest node always wins (deterministic execution)
  ├── Dishonest node gets slashed (collateral forfeited)
  └── Client who reported fraud receives compensation

  Security:
  ├── k redundant inferences → correct with prob 1 - f^k
  │   where f = fraction of Byzantine nodes
  └── EigenLayer operators for cryptoeconomic security

Component 7: Pulse Round System

PULSE ROUND (every 90 seconds)
══════════════════════════════

  ┌─────────────────────────────────┐
  │  All nodes participate          │
  │                                 │
  │  1. Broadcast presence signal   │
  │  2. Report capabilities         │
  │  3. Submit work receipts        │
  │                                 │
  │  Validation nodes verify:       │
  │  ├── Presence proofs            │
  │  ├── Work receipt integrity     │
  │  └── Capability claims          │
  │                                 │
  │  Points calculated:             │
  │  base_points = 10               │
  │  uptime_bonus = 1 + 0.2·ln(1 + t/12)
  │  capability_bonus = Σ weights   │
  │  work_points = tokens × rate    │
  │                                 │
  │  Total = (base × uptime_bonus   │
  │         × capability_bonus)     │
  │         + work_points           │
  └─────────────────────────────────┘

11 Core Mechanisms (Detailed)

Mechanism 1: Three-Layer Coordination Stack (Deep Dive)

The coordination stack is Hyperspace AGI's central architectural innovation. Each layer addresses a specific distributed systems challenge:

Layer 1: GossipSub — Real-Time Inspiration

GossipSub is libp2p's gossip-based pubsub protocol. Unlike flood-based gossip, it maintains a mesh topology where each node forwards messages only to a subset of peers (mesh members), with additional peers receiving metadata-only notifications (gossip peers) who can request full messages on demand.

GOSSIPSUB MESH TOPOLOGY
════════════════════════

  Node A ←──mesh──→ Node B ←──mesh──→ Node C
    │                  │                  │
    │gossip           │mesh              │gossip
    │                  │                  │
  Node D            Node E            Node F
    │                  │
    │mesh              │gossip
    │                  │
  Node G            Node H

  When Agent A discovers Kaiming init helps:
  1. A publishes to mesh peers (B, D, G)     [~50ms]
  2. B forwards to its mesh peers (C, E)     [~100ms]
  3. Gossip peers (F, H) receive IHAVE       [~200ms]
  4. F, H request full message via IWANT     [~500ms]
  5. Full propagation across network         [~1 second]

  Properties:
  ├── O(1) per-node message cost (bounded fan-out)
  ├── Resilient to node failures (mesh self-heals)
  ├── Peer scoring prevents spam
  └── Topic-based filtering (agents subscribe to domains)

Layer 2: Loro CRDTs — Convergent State

Loro is a high-performance CRDT library that provides convergent state without requiring consensus. In Hyperspace AGI, each research domain maintains a CRDT leaderboard:

CRDT CONVERGENCE EXAMPLE
═════════════════════════

  Time T=0: Three nodes, empty leaderboards

  Node X                Node Y                Node Z
  ┌──────┐              ┌──────┐              ┌──────┐
  │empty │              │empty │              │empty │
  └──────┘              └──────┘              └──────┘

  Time T=30s: X and Z discover results concurrently

  Node X                Node Y                Node Z
  ┌──────────┐          ┌──────┐              ┌──────────┐
  │ X: 0.95  │          │empty │              │ Z: 0.93  │
  └──────────┘          └──────┘              └──────────┘

  Time T=90s: Delta sync propagates (order-independent)

  Node X                Node Y                Node Z
  ┌──────────┐          ┌──────────┐          ┌──────────┐
  │ 1. Z:0.93│          │ 1. Z:0.93│          │ 1. Z:0.93│
  │ 2. X:0.95│          │ 2. X:0.95│          │ 2. X:0.95│
  └──────────┘          └──────────┘          └──────────┘

  All three nodes converge to identical state
  regardless of message delivery order.

  Properties:
  ├── Commutativity: merge(A, B) = merge(B, A)
  ├── Associativity: merge(A, merge(B, C)) = merge(merge(A, B), C)
  ├── Idempotency: merge(A, A) = A
  ├── Delta encoding: only send changes, not full state
  └── Zero cold start: new nodes receive full snapshot

Layer 3: GitHub Archive — Durable Reproducibility

The Git layer provides permanent, human-readable records. Every agent pushes to its own branch in hyperspaceai/agi:

GIT ARCHIVE STRUCTURE
═════════════════════

  hyperspaceai/agi
  ├── main
  │   ├── README.md
  │   └── snapshots/
  │       └── latest.json          ← Hourly CRDT snapshot
  │
  ├── agent/A-7f3e/
  │   ├── experiments/
  │   │   ├── exp_001.json         ← Kaiming init hypothesis
  │   │   ├── exp_002.json         ← RMSNorm variant
  │   │   └── exp_003.json         ← Combined approach
  │   └── papers/
  │       └── kaiming_rms_paper.md ← Synthesized findings
  │
  ├── agent/B-2e1a/
  │   └── experiments/
  │       ├── exp_001.json
  │       └── ...
  │
  └── agent/C-9a4b/
      └── ...

  Snapshot format (snapshots/latest.json):
  {
    "timestamp": "2026-03-15T12:00:00Z",
    "domains": {
      "ml": {
        "leaderboard": [...],
        "total_experiments": 847,
        "active_agents": 42
      },
      "search": {...},
      "finance": {...},
      "skills": {...},
      "causes": {...}
    }
  }

Mechanism 2: Distributed Response Cache (Mathematical Foundation)

The response cache exploits a fundamental property of deterministic LLM inference: SHA-256(model_id || params || prompt) produces the same hash for identical inputs, and the same model produces the same output for the same input (given deterministic sampling).

Cache Proof Verification:

CacheProof = {
    requestHash:  SHA-256(model_id || params || prompt),
    responseHash: SHA-256(response),
    proofHash:    SHA-256(requestHash || responseHash || metadata),
    computedAt:   ISO-8601 timestamp,
    signature:    Ed25519.sign(proofHash, node_private_key),
    ttl:          24 hours
}

Verification:
1. Recompute proofHash from (requestHash, responseHash, metadata)
2. Verify Ed25519 signature against node's public key
3. Check TTL has not expired
4. Accept response as authenticated

Popularity Amplification:

When a node fetches a cached response, it becomes a provider for that response. This creates a positive feedback loop where popular responses replicate naturally across the network—the most requested content becomes the most available, following the same principle as BitTorrent's piece selection.

Mechanism 3: KV Prefix Cache (Erasure Coding + KZG)

The KV prefix cache is the most technically sophisticated caching layer, combining information theory (erasure codes) with cryptographic commitments:

Reed-Solomon Erasure Coding:

ERASURE CODING FOR KV STATE
════════════════════════════

  Original KV state: 150 MB (Qwen 7B, 2K tokens)

  Reed-Solomon(k=32, n=64):
  ├── 32 data chunks   (each ~4.7 MB)
  ├── 32 parity chunks (each ~4.7 MB)
  └── Any 32 of 64 chunks → reconstruct full state

  Failure tolerance: up to 50% of chunks can be lost
  Distribution: chunks spread across 64 different peers
  Recovery: even if 32 peers go offline, state is recoverable

  Total redundancy: 2x (300 MB stored across network)
  Per-node storage: ~4.7 MB per chunk

KZG Polynomial Commitments:

KZG COMMITMENT SCHEME
══════════════════════

  Purpose: verify any chunk of a 600 MB KV cache
           with only a 48-byte commitment

  1. Encode KV state as polynomial P(x) over finite field
  2. Compute commitment C = [P(τ)]₁  (48 bytes, elliptic curve point)
  3. Broadcast C via GossipSub

  Verification (per chunk):
  1. Node requests chunk i from peer
  2. Peer provides chunk data + KZG opening proof πᵢ
  3. Node verifies: e(C - [vᵢ]₁, [1]₂) = e(πᵢ, [τ - i]₂)
     (bilinear pairing check, ~1ms)
  4. If valid: chunk is authentic → accept

  Properties:
  ├── 48 bytes to verify any chunk of arbitrarily large state
  ├── Computational hiding: cannot extract state from commitment
  ├── Binding: cannot produce valid proof for incorrect chunk
  └── Same scheme used in Ethereum danksharding (EIP-4844)

DAP (Data Availability Probing):

DAP PROBING PROTOCOL
════════════════════

  Goal: verify that peers actually store the chunks they claim

  Method: probes are indistinguishable from real requests

  1. Prober sends request for random chunk index
  2. Peer must respond with chunk + KZG proof
  3. Prober verifies proof

  Key insight: because probes look identical to real
  requests, a node cannot selectively respond to probes
  while withholding from real users.

  Failure consequence: reputation penalty
  └── Reduces future point earnings
  └── Eventually removed from provider lists

Mechanism 4: Fraud Proof Interactive Verification

The fraud proof system implements a bisection-based challenge protocol, analogous to optimistic rollup fraud proofs in blockchain systems:

FRAUD PROOF BISECTION PROTOCOL
═══════════════════════════════

  Neural network with 10M neurons (ordered):

  Round 1: Submit state roots for each 1M neurons
  ┌───┬───┬───┬───┬───┬───┬───┬───┬───┬───┐
  │1M │2M │3M │4M │5M │6M │7M │8M │9M │10M│
  └───┴───┴───┴───┴─×─┴───┴───┴───┴───┴───┘
                     ↑ Challenger: "root #5 is wrong"

  Round 2: Subdivide incorrect million into 100K groups
  ┌────┬────┬────┬────┬────┬────┬────┬────┬────┬────┐
  │4.1M│4.2M│4.3M│4.4M│4.5M│4.6M│4.7M│4.8M│4.9M│5.0M│
  └────┴────┴────┴──×─┴────┴────┴────┴────┴────┴────┘
                     ↑ Challenger: "root #4 is wrong"

  ... (continue bisecting) ...

  Round 7: Single neuron identified
  ┌────────────────────────────────────────────┐
  │  neuron_4,387,219                          │
  │  value = activation(Σ weight_j × input_j)  │
  │                                            │
  │  On-chain verification:                    │
  │  1. Retrieve weights from Merkle tree      │
  │  2. Compute sum + activation               │
  │  3. Compare with claimed value             │
  │  4. If mismatch → slash challenged node    │
  └────────────────────────────────────────────┘

  Complexity: O(log₁₀(10M)) = 7 rounds
  On-chain computation: one neuron evaluation

Mechanism 5: Threshold Signatures and Byzantine Fault Tolerance

THRESHOLD SIGNATURE SCHEME
══════════════════════════

  Setup:
  ├── n nodes, unified public key PK
  ├── Each node i has unique private key skᵢ
  └── Threshold k: any k nodes can produce valid signature

  Signing:
  1. Node i computes partial signature: ρᵢ ← tsign(skᵢ, m)
  2. Collect k partial signatures: {ρᵢ | i ∈ I, |I| = k}
  3. Combine: σ ← tcombine(m, {ρᵢ})
  4. Verify: tverify(PK, m, σ) → true

  Byzantine Fault Tolerance:
  ├── Up to f = |F| Byzantine nodes
  ├── k redundant inferences of same query
  ├── Same result across k nodes → correct with prob 1 - (f/n)^k
  ├── Different results → fraud proof triggered
  └── Honest nodes compensated, dishonest slashed

  EigenLayer Integration:
  └── Ethereum validators provide cryptoeconomic security
      Operators stake ETH → can be slashed for dishonesty
      Bootstraps trust without Hyperspace-native token

Mechanism 6: Emergent Cross-Agent Learning

The overnight experiment demonstrated emergent knowledge transfer:

EMERGENT LEARNING PROPAGATION
══════════════════════════════

  Timeline of Kaiming Initialization Discovery:

  T=0h:    Agent A-7f3 begins ML experiments
           Tries random init, Xavier init → baseline results

  T=2h:    Agent A-7f3 tries Kaiming initialization
           val_loss drops from 0.961 to 0.942 (1.98% improvement)

  T=2h:    A-7f3 broadcasts via GossipSub:
           "Kaiming init: val_loss 0.961 → 0.942"

  T=2.5h:  CRDT leaderboard updates across all nodes
           A-7f3 moves to rank #1 in ML domain

  T=3h:    Agent B-2e1 sees Kaiming init in leaderboard
           Adopts technique, combines with RMSNorm

  T=4h:    8 more agents adopt Kaiming init

  T=6h:    15 agents running Kaiming variants

  T=8h:    23 of 35 agents have adopted Kaiming init
           Multiple variants being explored simultaneously

  T=12h:   333 total experiments completed
           Knowledge has saturated the network

  This is analogous to how ideas spread through a
  research community, but compressed from months
  (conference cycles) to hours (gossip cycles).

12 Programming Language

Infrastructure Layer

The Hyperspace node infrastructure spans multiple languages:

Component Language Rationale
aios-cli Shell (Bash) Installer scripts, platform detection
HyperspaceZK TypeScript Zero-knowledge framework, browser integration
zkwasm-sdk TypeScript WebAssembly ZK SDK
Node client Likely Rust/Go Performance-critical P2P networking (inferred from libp2p usage)
Browser client TypeScript/JavaScript WebGPU integration, WebRTC transport

Research Layer (AGI)

The AGI repo itself functions as a living research artifact rather than a traditional codebase. Agent code runs on the Hyperspace node client, using the OpenAI-compatible API for LLM interactions.

Key Libraries and Dependencies

Library Purpose Used In
libp2p P2P networking (DHT, GossipSub, transports) All node types
Loro CRDT state management Leaderboard convergence
WebGPU/WebLLM Browser-based GPU inference Browser nodes
Reed-Solomon Erasure coding for KV cache Cache distribution
Ed25519 Digital signatures Cache proofs, identity
SHA-256 Content hashing Cache keys, Merkle trees
KZG Polynomial commitments KV chunk verification
EigenLayer Cryptoeconomic security Validator integration

Smart Contract Layer

The fraud proof and challenge system involves on-chain components:

ON-CHAIN COMPONENTS
════════════════════

  EVM-compatible smart contracts:
  ├── Fraud proof challenge/response
  ├── Collateral staking/slashing
  ├── Threshold signature verification
  └── EigenLayer operator registration

  Verification:
  ├── Merkle root validation
  ├── Single-neuron execution
  ├── Activation function computation
  └── Weight sum verification

Agent-Side Code

Research agents interact with the system through the OpenAI-compatible API, meaning agent logic can be written in any language with an HTTP client. The canonical pattern uses Python:

from openai import OpenAI

client = OpenAI(
    base_url="http://localhost:8080/v1",
    api_key="hyperspace"
)

def research_loop(domain: str, leaderboard: dict):
    """Continuous research cycle."""
    while True:
        # 1. Hypothesize (informed by leaderboard + gossip)
        hypothesis = client.chat.completions.create(
            model="qwen-7b",
            messages=[{
                "role": "system",
                "content": f"Domain: {domain}. Current SOTA: {leaderboard}"
            }, {
                "role": "user",
                "content": "Generate a novel hypothesis and experiment code"
            }]
        )

        # 2. Execute experiment
        result = run_experiment(hypothesis.choices[0].message.content)

        # 3. Share via gossip + CRDT + Git
        broadcast_result(result)

13 Memory Management

Distributed State Architecture

Hyperspace manages state across four distinct layers, each with different consistency, durability, and latency characteristics:

STATE MANAGEMENT HIERARCHY
══════════════════════════

  ┌─────────────────────────────────────────────────┐
  │  Layer 4: PERMANENT ARCHIVE (Git)               │
  │  ├── Per-agent branches in hyperspaceai/agi     │
  │  ├── Hourly snapshots (snapshots/latest.json)   │
  │  ├── Immutable once committed                   │
  │  ├── Human-readable (JSON, Markdown)            │
  │  └── Latency: ~5 minutes                        │
  ├─────────────────────────────────────────────────┤
  │  Layer 3: CONVERGENT STATE (Loro CRDTs)         │
  │  ├── Per-domain leaderboards                    │
  │  ├── Eventually consistent (no coordination)    │
  │  ├── Delta-only sync between peers              │
  │  ├── Full snapshot on new node join              │
  │  └── Latency: ~2 minutes                        │
  ├─────────────────────────────────────────────────┤
  │  Layer 2: EPHEMERAL CACHE (DHT + Local)         │
  │  ├── Response cache: SQLite + DHT providers      │
  │  ├── KV prefix cache: erasure-coded chunks      │
  │  ├── TTL-based expiration (24h for responses)   │
  │  ├── Popularity amplification (fetcher → provider)│
  │  └── Latency: 1ms (local) to 200ms (DHT)       │
  ├─────────────────────────────────────────────────┤
  │  Layer 1: REAL-TIME GOSSIP (GossipSub)          │
  │  ├── Fire-and-forget messages                   │
  │  ├── Best-effort delivery                       │
  │  ├── Not persisted (consumed or dropped)        │
  │  ├── Topic-based subscription                   │
  │  └── Latency: ~1 second                         │
  └─────────────────────────────────────────────────┘

Node-Local State

Each node maintains several local data stores:

Store Technology Contents Persistence
Response Cache SQLite SHA-256 indexed responses Persistent, TTL-evicted
KV Prefix State In-VRAM / on-disk Attention state for active models Ephemeral (VRAM) / cached (disk)
CRDT Replica Loro in-memory Full leaderboard state per domain Reconstructed from peers on restart
Peer Table k-bucket lists Known peers by XOR distance Persistent, refreshed
Identity Keystore Ed25519 keypair + NodeAddress Persistent
Work Receipts Local log Task completion proofs Persistent until submitted

Memory Capability: Distributed Vector Store

Nodes with the Memory capability (+5% points) operate a distributed vector store with replication:

DISTRIBUTED VECTOR STORE
═════════════════════════

  Agent queries:
  "Find experiments similar to 'Kaiming init for transformers'"
      │
      ▼
  ┌──────────────────────┐
  │  Local Embedding     │  MiniLM-L6-v2 (CPU)
  │  query → vector      │  384-dimensional
  └──────────┬───────────┘
             │
             ▼
  ┌──────────────────────┐
  │  Distributed Search  │
  │  ├── Local index     │  Nearest neighbors
  │  ├── DHT routing     │  Find memory nodes
  │  └── Aggregate       │  Merge results
  └──────────────────────┘

  Properties:
  ├── Replicated across memory-capable nodes
  ├── Semantic search across all agent experiments
  ├── Cross-domain knowledge retrieval
  └── Used by research loop for hypothesis generation

Cache Hit Rate Dynamics

The cache system exhibits emergent efficiency as network size grows:

CACHE EFFICIENCY vs. NETWORK SIZE
═════════════════════════════════

  Hit Rate
  100% ┤
       │                              ╭──────── theoretical max
   90% ┤                          ╭───╯
       │                      ╭───╯
   80% ┤                  ╭───╯
       │              ╭───╯
   70% ┤          ╭───╯
       │      ╭───╯           Zipf's law: popular queries
   60% ┤  ╭───╯               dominate, cache fills itself
       │╭─╯                   with what matters most
   50% ┤╯
       │
   40% ┤
       │
   30% ┤
       └──────┬──────┬──────┬──────┬──────
            10K   100K    1M    10M   100M  nodes

  Key insight: sub-linear cost scaling
  ├── Linear systems: 2x users = 2x cost
  └── Hyperspace:     2x users = higher hit rate = lower cost/user

Garbage Collection and TTL Management

  • Response cache: 24-hour TTL on Ed25519-signed cache proofs; entries evicted after TTL or LRU when local storage fills
  • KV prefix cache: ephemeral by nature; chunks evicted when VRAM is needed for active inference; on-disk fallback for high-value prefixes
  • CRDT state: grows monotonically (append-only leaderboards); compaction via snapshot + truncation of superseded entries
  • Gossip messages: consumed on receipt, not persisted; mesh topology ensures delivery without storage overhead

14 Continued Learning

Cross-Agent Knowledge Transfer

Hyperspace AGI's most distinctive continued learning mechanism is its gossip-mediated knowledge transfer. Unlike centralized systems where a knowledge base must be explicitly maintained and queried, Hyperspace leverages the P2P gossip protocol for organic knowledge propagation:

KNOWLEDGE PROPAGATION MODEL
════════════════════════════

  Traditional (Centralized):
  Agent → Central KB → Agent
  ├── Single point of failure
  ├── Bottleneck at scale
  └── Curated (editorial overhead)

  Hyperspace (Decentralized):
  Agent A ←gossip→ Agent B ←gossip→ Agent C
  ├── No single point of failure
  ├── Scales with network size
  ├── Self-curating (leaderboard ranks)
  └── Emergent (no editorial overhead)

Multi-Domain Cross-Pollination

Because agents operate across 5 domains with shared gossip infrastructure, discoveries in one domain can inform hypotheses in another:

CROSS-DOMAIN KNOWLEDGE FLOW
════════════════════════════

  ML Domain                    Search Domain
  ┌─────────────────┐         ┌─────────────────┐
  │ "RMSNorm helps   │         │ "BM25 + neural   │
  │  transformer      │ ─────→ │  reranker with    │
  │  training"        │         │  normalization    │
  └─────────────────┘         │  improvements"    │
                              └─────────────────┘

  Skills Domain                Financial Domain
  ┌─────────────────┐         ┌─────────────────┐
  │ "Improved WASM   │         │ "Better data      │
  │  data extraction" │ ─────→ │  preprocessing    │
  │                   │         │  for backtesting" │
  └─────────────────┘         └─────────────────┘

  The Causes meta-domain explicitly facilitates
  cross-pollination by running sub-causes across
  all other domains.

Peer Review and Quality Filtering

The 6-stage research loop includes a peer review mechanism that serves as a quality filter for continued learning:

  1. Accumulation — Agent runs N experiments, accumulates results
  2. Synthesis — Agent writes a research paper summarizing findings
  3. Peer Review — Other agents read, critique, and score papers (1-10 scale)
  4. Threshold — Papers scoring 8+ are flagged as "breakthroughs"
  5. Feedback — Breakthrough papers feed back into hypothesis generation
  6. Evolution — Future agents benefit from curated, high-quality findings

This creates a natural quality gradient: raw experimental results propagate via gossip (fast, unfiltered), while synthesized knowledge propagates via peer review (slower, quality-filtered).

Leaderboard-Driven Curriculum

The CRDT leaderboard implicitly creates a curriculum for new agents:

NEW AGENT ONBOARDING
════════════════════

  1. Agent joins network
  2. Receives full CRDT leaderboard snapshot (zero cold start)
  3. Observes:
     ├── Current SOTA techniques (top of leaderboard)
     ├── Active research directions (recent entries)
     ├── Tried-and-failed approaches (low-ranking entries)
     └── Promising unexplored regions (gaps in leaderboard)
  4. Agent formulates initial hypothesis informed by
     the entire history of network research
  5. Begins contributing immediately (no warmup period)

DiLoCo Continuous Model Improvement

When multiple agents collaboratively train models via DiLoCo, the resulting models improve continuously:

  • Each agent trains locally for H steps, then shares compressed weight deltas
  • Deltas are averaged across participating agents
  • The averaged model captures diverse training strategies
  • Automatic fallback to solo training if no peers are available
  • FP16 gradient all-reduce maintains training quality

Network-Level Learning Curves

At the network level, Hyperspace AGI exhibits a learning dynamic where:

  1. Early phase: Agents independently explore the search space, discovering low-hanging fruit
  2. Gossip phase: Successful techniques propagate, agents converge on productive directions
  3. Refinement phase: With baseline techniques established, agents explore variations and combinations
  4. Plateau: Diminishing returns as obvious improvements are exhausted
  5. Breakthrough: Occasional paradigm shifts when an agent discovers a fundamentally new approach

This mirrors the dynamics of human research communities but operates on a compressed timescale (hours vs. months).

Comparison with Other Continued Learning Approaches

Approach Hyperspace AGI FunSearch (DeepMind) OpenELM Autoresearch
Learning unit Agent population Program population Map population Single agent
Transfer mechanism P2P gossip + CRDTs Prompt sampling Curiosity-driven None
Cross-domain Yes (5 domains) No (single problem) Yes (evolved) No
Persistence Git + CRDT Database Archive Local logs
Quality filter Peer review (8+ score) Score-based sampling Fitness function Agent judgment
Cold start Zero (CRDT snapshot) Seed programs Seed population From scratch
Scale Thousands of agents Hundreds of workers Dozens One

15 Applications

Current Applications (Production)

1. Machine Learning Architecture Search

Agents autonomously explore neural network architectures for language modeling on astrophysics papers, optimizing val_loss. This directly extends Karpathy's autoresearch to distributed multi-agent execution.

2. Search Engine Ranking Evolution

Agents evolve BM25 + neural reranker pipelines, optimizing NDCG@10 for web search quality. This combines traditional IR techniques with learned components.

3. Financial Strategy Discovery

Agents backtest S&P 500 monthly-rebalance strategies, optimizing Sharpe ratio. Autonomous exploration of portfolio construction techniques.

4. WASM Skill Forging

Agents develop WebAssembly skills for web scraping, data extraction, and parsing—evaluated by test_pass_rate. This is a self-improvement loop where agents build tools for other agents.

5. Causes Research (Meta-Domain)

A meta-domain that encompasses sub-causes: search ranking improvement, literature analysis, skill forge enhancement, infrastructure development, and data curation.

Infrastructure Applications

6. Decentralized AI Inference

The 2M+ node network serves as a decentralized alternative to centralized AI providers. Any tool compatible with the OpenAI API can use Hyperspace as a backend—replacing API calls to OpenAI/Anthropic/Google with calls to localhost:8080/v1.

7. Distributed Inference Caching

The three-layer cache (response + KV prefix + full inference) eliminates redundant computation at network scale. At 1M nodes, an estimated 70-80% of inference requests are served from cache.

8. Privacy-Preserving AI Access

The proxy capability enables agents to access resources through residential IP proxies, and the fully decentralized architecture means no single entity controls or logs inference requests.

Potential Future Applications

9. Distributed Model Training at Scale

If DiLoCo collaborative training matures, the 2M+ node network could become a distributed training platform for open-weight models—competing with centralized training clusters.

10. Autonomous Scientific Discovery

The multi-domain research framework is domain-agnostic. New research domains can be added by defining a metric and evaluation protocol. Potential domains include: - Drug discovery (binding affinity optimization) - Materials science (property prediction) - Theorem proving (proof success rate) - Code generation (benchmark pass rate)

11. Decentralized AI Research Community

The peer review mechanism (agents scoring papers 8+) could evolve into a decentralized research community where agents produce, review, and build upon each other's work—an autonomous version of the academic publishing process.

12. Federated Knowledge Base

The combination of distributed vector store, CRDT state, and gossip-based knowledge propagation could evolve into a federated knowledge base that grows organically from network research activity.

Significance for the AI Discovery Ecosystem

Hyperspace represents a qualitatively different approach to AI-driven discovery:

POSITIONING IN THE DISCOVERY LANDSCAPE
═══════════════════════════════════════

  Single Agent                          Multi-Agent
  Low Scale                             High Scale
      │                                     │
      ▼                                     ▼
  ┌──────────┐  ┌──────────┐  ┌──────────┐  ┌──────────────┐
  │Karpathy  │  │FunSearch │  │OpenELM   │  │ HYPERSPACE   │
  │autoresrch│  │(DeepMind)│  │          │  │    AGI       │
  │          │  │          │  │          │  │              │
  │1 agent   │  │100s of   │  │Diverse   │  │ 1000s of     │
  │1 GPU     │  │workers   │  │pop'ns    │  │ P2P agents   │
  │1 domain  │  │Central   │  │MAP-Elites│  │ 5 domains    │
  │No sharing│  │database  │  │Archive   │  │ Gossip+CRDT  │
  │          │  │1 problem │  │          │  │ Git archive  │
  └──────────┘  └──────────┘  └──────────┘  │ Incentivized │
                                             │ 2M+ nodes    │
      Centralized ─────────────────────────  │ Decentralized│
                                             └──────────────┘

  Key differentiation:
  ├── SCALE: 2M+ nodes vs. single GPU or controlled cluster
  ├── COORDINATION: gossip + CRDTs vs. central DB/none
  ├── DOMAINS: 5 simultaneous vs. single problem
  ├── INCENTIVES: economic alignment vs. none
  ├── ACCESS: browser tab vs. H100 cluster
  └── PERSISTENCE: Git + CRDT vs. local logs

Open Questions for the Field

  1. Quality vs. Quantity — Does distributed autoresearch produce genuine insights, or just more parameter sweeps? The Kaiming initialization discovery is encouraging but anecdotal.

  2. Incentive Alignment — Does the points economy attract researchers (who optimize for discoveries) or node farmers (who optimize for points)? The +12% research weight is a design-level signal, but real-world behavior may diverge.

  3. Emergent Intelligence — Does the gossip-mediated knowledge transfer produce genuinely emergent intelligence (insights no single agent would find alone), or is it merely parallel search with communication?

  4. Scaling Laws — How does research quality scale with network size? Linear? Sub-linear? Is there a critical mass beyond which the network becomes self-sustaining?

  5. Domain Transfer — Can techniques discovered in one domain (e.g., ML architecture insights) genuinely transfer to another (e.g., financial strategy design), or are the domains too different for cross-pollination?

  6. Verification at Scale — As the network grows, can the fraud proof and verification systems maintain integrity, or do adversarial strategies emerge that exploit the decentralized trust model?

These questions position Hyperspace AGI as both a production system and a research experiment—a living laboratory for studying distributed AI-driven discovery at scale.


Analysis based on: Hyperspace whitepaper (hyperspace.computer/bittorrent-for-ai.pdf), GitHub repositories (hyperspaceai/agi, hyperspaceai/hyperspace-node, hyperspaceai/aios-cli), distributed cache specification (cache.hyper.space), Ry Walker Research analysis (rywalker.com/research/hyperspace-agi), Khan et al. 2025 "A Gossip-Enhanced Communication Substrate for Agentic AI" (arXiv:2512.03285), and network dashboard data.