Gonka Broker
The easiest way to use open‑source LLMs

Inference of open‑source LLMs
at the lowest possible cost

Gonka Broker gives you an OpenAI‑compatible API to the world's largest decentralized GPU network. Accounts, API keys, card payments — no wallets or tokens.

Developers choose Gonka Broker

We remove crypto infrastructure barriers. Leave only API and cheap inference.

OpenAI‑compatible API

Just change the base_url in any OpenAI SDK. Supports Python, JS, and other clients.

💳

Billing in USD

Top up with familiar payment methods, receive simple invoices. No tokens, wallets, or on‑chain steps.

🔑

API keys & accounts

Create separate keys for different apps and environments. Full access control.

🌐

Largest decentralized GPU network

Access the power of Gonka's global GPU infrastructure.

Start in 3 steps

Instant access without diving into the decentralized network.

1

Create an account

Sign up with email or Google. No wallets, seed phrases, or crypto.

2

Generate an API key

In the dashboard, create a key for your application or environment.

3

Send requests

Choose your integration mode below and start inferring.

Two ways to connect

Both give you native USD billing, low prices, open models, and the power of the world's largest decentralized GPU network. Choose the one that fits your workflow.

🛠️

Direct Mode (SDK / Developer)

Maximum privacy & control. Your software talks directly to the Gonka network — we never see your prompts or completions.

  • Direct peer-to-peer inference
  • Data never passes through our servers
  • Full control over endpoint selection
  • Requires a few lines of integration
import OpenAI from "openai";
import { resolveAndSelectEndpoint, gonkaFetch } from "gonka-openai";

const { selected } = await resolveAndSelectEndpoint({
  sourceUrl: "https://node3.gonka.ai",
});

const fetch = gonkaFetch({
  gonkaPrivateKey: process.env.YOUR_PRIVATE_API_KEY,
  selectedEndpoint: selected,
});

const client = new OpenAI({
  apiKey: "any-string",
  baseURL: selected.url,
  fetch,
});

const response = await client.chat.completions.create({
  model: "Qwen/Qwen3-32B-FP8",
  messages: [{ role: "user", content: "Hello!" }],
});
    
🔁

Proxy Mode (3rd party apps)

Instant compatibility. Plug any OpenAI‑compatible tool into Gonka — just replace the base_url and API key.

  • Works with LangChain, LlamaIndex, ChatGPT‑like UIs
  • Zero code changes required
  • All requests are forwarded (we do not store them)
  • Ideal for existing projects and no‑code integrations
from openai import OpenAI

client = OpenAI(
    base_url="https://proxy.gonkabroker.com/v1",  
    api_key="your_broker_key"  
)

response = client.chat.completions.create(
    model="Qwen/Qwen3-32B-FP8",
    messages=[{"role": "user", "content": "Hello!"}]
)
    

Both modes give you access to the same low prices, open models, and the Gonka network. The difference is where the request is routed.

Inference pricing

Market-based pricing in a decentralized network is often lower than centralized platforms. *Sample prices per 1M tokens.

Model / Provider Input (per 1M tokens) Output (per 1M tokens)
OpenAI GPT-5 $1.25 $10.00
Claude 4.5 Sonnet $3.00 $15.00
Gonka (Qwen/Qwen3-235B) $0.35 $0.35

*Prices on the Gonka network may fluctuate based on GPU supply and demand. Shown for illustrative purposes.

Accepted payment methods

Visa Mastercard Union Pay Apple Pay Google Pay Link

and others

For developers

Works with your favorite tools

Whether you choose Direct Mode for maximum privacy or Proxy Mode for instant compatibility, you get a seamless developer experience.

  • Direct Mode

    use our gonka-openai library to connect directly to network nodes. Your data stays between you and the GPU providers.

  • Proxy Mode

    just point your existing OpenAI client to https://proxy.gonkabroker.com/v1 and start using Gonka models.

Quick start (Proxy Mode)

  • 1. Sign up – account in 30 seconds
  • 2. Create a key in the dashboard
  • 3. Change base_url to `https://proxy.gonkabroker.com/v1` and start
🔒

We don't store your data

Gonka Broker only forwards requests to the decentralized network. Prompts and completions are not logged or stored. Your code and data remain yours.

Start using cheap inference now

No crypto wallets, no complexity. Just API and cheap LLMs.

Sign up and start

Already have an account? Sign in