Skip to content

What is Inference Grid?

The Inference Grid is a decentralized network that enables distributed AI model inference across a global network of independent providers. Built on Lightning Network for seamless micropayments, it provides censorship-resistant access to both open source and proprietary language models.

The platform is permissionless, allowing anyone to participate as a provider while users can easily integrate it as a drop-in replacement for traditional AI providers like OpenAI or Anthropic.

Use Cases

  • Cost savings: Use the Inference Grid as a drop-in replacement for OpenAI and reduce costs with our optimized model routing.
  • Pay-per-call APIs: Build AI-powered APIs where users only pay for the compute they use, with no upfront costs or subscriptions
  • Uncensored Chat Apps: Create chat applications that can access both open source and proprietary models without content restrictions
  • Research Tools: Develop research assistants that can leverage multiple models simultaneously for better results
  • AI Marketplaces: Build marketplaces where users can access different models through a unified interface
  • Privacy-Focused Apps: Create applications that don't require users to share data with centralized providers