0G Compute
Use cheap, TEE-verified AI models from the 0G Compute Network as OpenClaw providers. Discover available models and compare pricing vs OpenRouter, verify prov...
Description
name: 0g-compute description: "Use cheap, TEE-verified AI models from the 0G Compute Network as OpenClaw providers. Discover available models and compare pricing vs OpenRouter, verify provider integrity via hardware attestation (Intel TDX), manage your 0G wallet and sub-accounts, and configure models in OpenClaw with one workflow. Supports DeepSeek, GLM-5, Qwen, and other models available on the 0G marketplace." metadata: {"openclaw":{"requires":{"bins":["0g-compute-cli"]},"install":[{"id":"node","kind":"node","package":"@0glabs/0g-serving-broker","bins":["0g-compute-cli"],"label":"Install 0G Compute CLI"}]}}
0G Compute Network
Interface with the 0G Compute Network — a decentralized AI inference marketplace with TEE-verified model integrity.
Prerequisites
0g-compute-cliinstalled:npm i -g @0glabs/0g-serving-broker- Note: npm package name is
@0glabs/0g-serving-broker(the old@0glabs/0g-compute-clipackage name no longer resolves), but the binary command is still0g-compute-cli.
- Note: npm package name is
- Wallet funded with 0G tokens
- Logged in: avoid passing key on the command line when possible (it can be visible in shell history/process list). Prefer reading from a protected prompt/env:
read -s OG_PK; 0g-compute-cli login --private-key "$OG_PK"; unset OG_PK
- Network configured:
0g-compute-cli setup-network
Core Workflows
1. Discover Models
# List all providers with models, prices, verifiability
0g-compute-cli inference list-providers
# Detailed view with health/uptime metrics
0g-compute-cli inference list-providers-detail
# Include providers without valid TEE signer
0g-compute-cli inference list-providers --include-invalid
Filter output by model name, price, health status, and TeeML support (models running in Trusted Execution Environment).
2. Verify Provider Integrity
Always verify before trusting a new provider. TEE verification ensures the model runs in a secure enclave with hardware-attested integrity.
# Full TEE attestation check
0g-compute-cli inference verify --provider <address>
# Download raw attestation data
0g-compute-cli inference download-report --provider <address> --output report.json
The verify command checks:
- TEE signer address matches contract
- Docker Compose hash integrity
- DStack TEE (Intel TDX) attestation
3. Wallet & Balance
# Account overview
0g-compute-cli get-account
# Per-provider sub-account balance
0g-compute-cli get-sub-account --provider <address> --service inference
# Fund operations
0g-compute-cli deposit --amount <0G> # To main account
0g-compute-cli transfer-fund --provider <addr> --amount <0G> --service inference # To sub-account
0g-compute-cli retrieve-fund --service inference # From sub-accounts
0g-compute-cli refund --amount <0G> # Withdraw to wallet
Important: Inference calls fail if sub-account balance is depleted. Monitor balances regularly.
4. Configure OpenClaw Provider
Get API key from a verified provider (interactive — prompts for expiration):
0g-compute-cli inference get-secret --provider <address>
Add to openclaw.json:
"providers": {
"0g-<model-name>": {
"baseUrl": "<provider-url>/v1/proxy",
"apiKey": "<secret>",
"api": "openai-completions",
"models": [{ "id": "<model-id>", "name": "<display-name>" }]
}
}
Register in agents.defaults.models with an alias. Set cost: 0 since billing is on-chain.
See references/openclaw-config.md for complete setup guide.
5. Price Comparison (0G vs OpenRouter)
Compare 0G pricing against OpenRouter for the same models:
scripts/0g-price-compare.sh
No API keys needed — uses public endpoints:
- CoinGecko for 0G token → USD price
- OpenRouter
/api/v1/modelsfor model pricing - 0G CLI for provider pricing
Shows side-by-side USD/1M tokens with savings percentage. Set OG_TOKEN_PRICE_USD env var to override CoinGecko price.
6. Status Check
# Login status & wallet
0g-compute-cli status
# Current network (mainnet/testnet)
0g-compute-cli show-network
Safety Guidelines
- Always run
inference verifyon new providers before use - Check provider health/uptime before relying on them
- Monitor sub-account balances — depleted funds cause inference failures
- Private key stored in
~/.0g-compute-cli/config.json— never expose this file
Reference Documentation
- CLI Reference — Complete command reference with all flags and examples
- OpenClaw Config — Step-by-step provider setup guide
Reviews (0)
No reviews yet. Be the first to review!
Comments (0)
No comments yet. Be the first to share your thoughts!