๐Ÿ’ 
bitGPT Docs
  • Introduction
    • ๐ŸŒWhat is bitGPT Network?
    • ๐Ÿค–How does it all work?
  • Infrastructure
    • โ˜ธ๏ธCore Architecture: Hub-and-Spoke
    • ๐Ÿ“ฆDistributed Storage Intelligence
    • ๐Ÿ—ž๏ธProxy Layer
    • ๐Ÿ›ตSmall Language Models
    • ๐Ÿ”งPrivacy-Preserving Fine-tuning
  • Getting Started
    • ๐Ÿ–ฅ๏ธYour wallet co-pilot: Terminal
    • ๐Ÿ“กWhat to expect
    • ๐Ÿ”ฎUp next: Open Registry
  • Ecosystem
    • ๐ŸŽŽPartnerships
    • โš™๏ธIntegrations
    • ๐Ÿ—บ๏ธAvailable Networks
  • Resources
    • ๐Ÿ“„Whitepaper
    • ๐Ÿ”ฌInsights
    • ๐ŸกExecutive team
Powered by GitBook
On this page
  1. Infrastructure

Core Architecture: Hub-and-Spoke

PreviousHow does it all work?NextDistributed Storage Intelligence

Last updated 2 months ago

The bitGPT Network is built on a hub-and-spoke architecture, a design that elegantly balances centralized power with decentralized efficiency. Imagine a bustling city where the central hub acts as the heartbeat, managing heavy lifting tasks like model training and data processing. The spokes, spread out across the city, are the edge devices โ€” your smartphones, computers, tablets โ€” handling local AI operations. This setup ensures that while the hub takes care of the intensive work, your device keeps your data private and processes AI tasks locally, reducing latency and enhancing security.

Hubs are the nerve centers of the network, comprising sponsors, curators, proxy operators, and engineers. They form the infrastructural and governance backbone, orchestrating resource allocation and ensuring the network operates smoothly. Spokes, on the other hand, are the edge devices where AI agents are deployed, powered by SLMs. These spokes deliver a self-sovereign AI experience, allowing users to maintain control over their data and computations.

This architecture isn't just about efficiency; it's a strategic move towards decentralization. By distributing computational tasks between hubs and spokes, the network avoids the pitfalls of centralization, like single points of failure and data breaches. Itโ€™s a symphony where each part plays its role, ensuring the network is both powerful and resilient.

The hub-and-spoke topology delivers substantive benefits that enhance both individual user experience and broader ecosystem sustainability. From a computational efficiency perspective, users experience markedly reduced latency for routine AI operations through edge processing on spokes. The architecture also catalyzes powerful network effects through its unique structure. Each additional spoke node contributes to the networkโ€™s collective intelligence through privacy-preserving federated learning, enhancing model performance without compromising individual data sovereignty.

The hub coordination layer facilitates efficient resource discovery and allocation, optimizing utility for both resource providers and consumers. These dynamics create a self-reinforcing cycle where enhanced user experience strengthens network effects and amplifies system utility, addressing both immediate user requirements and long-term ecosystem sustainability imperatives.

โ˜ธ๏ธ
Network Architecture