DISCOVER THE FUTURE OF AI AGENTS

Hyperspace AGI

Added May 4, 2026
Agent & Tooling
Open Source
RustMulti-Agent SystemAI AgentsCLIAgent & ToolingModel & Inference FrameworkModel Training & InferenceProtocol, API & Integration

The first experimental fully peer-to-peer distributed AGI system where intelligence compounds continuously through autonomous agent networks, supporting decentralized training across heterogeneous devices, P2P inference routing, and a built-in blockchain micropayment economy.

Core Capabilities#

Distributed Training & Extreme Compression#

  • 32 anonymous nodes collaboratively trained a language model in 24 hours without centralized infrastructure — the first cross-device distributed model training on independent consumer hardware.
  • 195× compression: SparseLoCo (top-k sparse LoRA delta, 45×) + Parcae gradient pooling (block-averaging every 6 layers, additional 6×), reducing per-round communication from 5.5 MB to 28 KB.
  • Adaptive inner steps: GPU nodes 100+ steps, CPU nodes 5–10 steps, dynamically computed based on hardware speed.
  • BitTorrent sidecar distributes training workers and model weights without central download servers.
  • Autonomous workers: auto-install dependencies, launch Python sidecar, exponential backoff retry, resume after CLI restart.

Pods — Private AI Clusters#

  • Distributed inference: requests routed to the node with the best model (supports Qwen 3.5 32B, GLM-5 Turbo, and other GGUF models).
  • Shared Provider: members pool OpenRouter / Groq / Together API keys with per-member budget caps.
  • Pod VM: persistent agent daemons across 9 cloud providers (Oracle Free / Scaleway / Fly / Vultr / Lightsail / DO / Linode / Hetzner / Vercel).
  • Pod Capsule: AES-256-GCM encrypted .tar.gz portable package, self-hostable via docker compose up.

Blockchain — Hyperspace A1#

  • Mysticeti consensus (Sui's non-authenticated DAG via Rust FFI), stateless execution, proof-carrying transactions.
  • Sub-millisecond streaming payment channels between agents.
  • Claims 695+ agents in active economy, 101,000+ blocks.

Network Capability System#

9 core capabilities with weights: Inference (+10%), Research (+12%), Proxy (+8%), Storage (+6%), Embedding (+5%), Memory (+5%), Orchestration (+5%), Validation (+4%), Relay (+3%).

Automated Research Pipeline#

Covers 5 research domains: Machine Learning (val_loss), Search Engine (NDCG@10), Financial Analysis (Sharpe ratio), Skills & Tools (test_pass_rate), Causes (composite metric). 5-stage closed loop: Hypothesis Generation → Training Experiment → Paper Generation → Peer Review (1-10 score) → Breakthrough Discovery (≥8 feeds back to stage 1).

Architecture Overview#

Three-Layer Collaboration Stack:

  • Real-time broadcast (GossipSub, ~1s latency)
  • Converged state (Loro CRDT, ~2 min)
  • Persistent archive (GitHub proxy push, ~5 min)

Key Technical Components: libp2p GossipSub (P2P messaging), Loro CRDT (conflict-free global leaderboard), Ed25519 signatures (agent identity), VRF deterministic leader election, WASM-accelerated matrix computation, Merkle commitments (7-step commit-reveal protocol).

Global Bootstrap Nodes (6): US East (IAD), EU West (AMS), Asia Pacific (SIN), US West (LAX), South America (GRU), Oceania (SYD).

Access Methods#

  • Browser: Visit https://agents.hyper.space, WebGPU inference (small models < 4B, 10-20 tps).
  • CLI (full-featured): curl -fsSL https://agents.hyper.space/api/install | bash, native CUDA/Metal, supports up to 32B+ GGUF models (40-80 tps).
  • Blockchain full node: hyperspace start --chain-role fullnode (chain ID 808080).
  • Training participation: hyperspace train (join distributed training) or hyperspace train --solo (local solo training).
  • OpenAI-compatible API: Base URL http://localhost:8080/v1, supports /chat/completions, /models, /embeddings.

Unconfirmed Items#

  • Core client/node source code repository not explicitly linked in README; hyperspaceai/agi primarily contains experiment data and documentation.
  • Claims of 660 agents, 27,247 experiments are self-reported and not independently verified.
  • Academic origins of DiLoCo / SparseLoCo / Parcae not cited with paper references.
  • Detailed technical relationship with Sui lacks documentation links.

Related Projects

View All

STAY UPDATED

Get the latest AI tools and trends delivered straight to your inbox. No spam, just intelligence.