Your GPU is idle.
Someone's agent needs it.
Share spare compute with the network. Use Qwen3 8B for free in Claude Code, Continue, and other agentic tools. No tokens, no fees — just a community.
Share spare compute with the network. Use Qwen3 8B for free in Claude Code, Continue, and other agentic tools. No tokens, no fees — just a community.
Grab the dollama CLI. Run as a user to consume inference, as a contributor to share compute, or both to do it all.
The coordinator matches user requests with available contributor nodes. All traffic routes through our relay — your files and context never leave your machine.
No tokens, no billing, no blockchain. Contributors earn priority in a leaderboard. The more you give, the faster your own requests get served.
Point your coding tools at dollama's local proxy. It forwards your LLM requests to the network and streams responses back. Your code and context stay local — only the inference is offloaded.
$ dollama connect
If you have Ollama running with Qwen3 8B pulled, donate your idle cycles. The CLI registers your node with the coordinator and handles incoming inference requests automatically.
$ dollama serve
curl -fsSL https://dollama.net/install.sh | sh
Requires Ollama installed with qwen3:8b pulled if you plan to contribute.
irm https://dollama.net/install.ps1 | iex
Requires Ollama installed with qwen3:8b pulled if you plan to contribute.
Run dollama connect to start the local proxy on port 11411. This exposes an Anthropic-compatible Messages API endpoint.
Copy the JSON config into ~/.claude/settings.json. This sets the base URL and auth token so Claude Code routes requests through the dollama network.
That's it. Claude Code will use the network model for inference. Your main Claude model stays unchanged — only the configured model routes through dollama.
Run dollama both instead — it starts the proxy and registers your machine as a contributor node in one command.
{
"env": {
"ANTHROPIC_BASE_URL": "http://localhost:11411",
"ANTHROPIC_AUTH_TOKEN": "olm_your_token_here"
},
"model": "network:qwen3:8b"
}
export ANTHROPIC_BASE_URL=http://localhost:11411
export ANTHROPIC_AUTH_TOKEN=olm_your_token_here
ollama pull qwen3:8b
| Rank | Contributor | Compute |
|---|---|---|
| 1 | gpu-farm-42 | 482,100 tok |
| 2 | inference-node-alpha | 310,850 tok |
| 3 | llm-volunteer-99 | 198,420 tok |
| 4 | weekend-contributor | 87,300 tok |
| 5 | idle-gpu-donor | 45,670 tok |
qwen3:8b pulled. Then just run dollama serve. The CLI handles registration, heartbeats, and inference routing automatically. Any machine that can run Ollama can contribute — dedicated GPUs are great, but even a modern CPU works.localhost:11411. Any tool that supports a custom Anthropic-compatible base URL can use it — Continue, Aider, or your own scripts. We're focused on agentic coding tools but the API is standard.