AI compute is centralized. A handful of providers control access, pricing, and availability. Today we are launching ACI (AI Compute Infrastructure), a decentralized network for trustless, verifiable AI inference.
The Centralization Problem
Current AI infrastructure:
- Few providers: OpenAI, Anthropic, Google control most capacity
- Opaque pricing: Costs change without notice
- Single points of failure: Outages affect everyone
- No verification: Trust that model ran correctly
- Geographic restrictions: Access varies by region
This centralization creates risk for AI-dependent businesses.
What is ACI
ACI is a blockchain-based network where:
- Compute providers offer GPU capacity
- Consumers pay for inference
- Validators verify computation
- Token holders govern the protocol
┌─────────────────────────────────────────────────┐
│ ACI Network │
│ │
│ ┌──────────┐ ┌──────────┐ ┌──────────┐ │
│ │ Provider │ │ Provider │ │ Provider │ │
│ │ GPU │ │ GPU │ │ GPU │ │
│ └────┬─────┘ └────┬─────┘ └────┬─────┘ │
│ │ │ │ │
│ └─────────────┼─────────────┘ │
│ │ │
│ ┌──────┴──────┐ │
│ │ Consensus │ │
│ │ Layer │ │
│ └──────┬──────┘ │
│ │ │
│ ┌──────┴──────┐ │
│ │ Consumers │ │
│ └─────────────┘ │
│ │
└─────────────────────────────────────────────────┘How It Works
Job Submission
Consumer submits inference job:
import { ACI } from '@hanzo/aci';
const aci = new ACI({ wallet: myWallet });
const result = await aci.inference({
model: 'llama-70b',
prompt: 'Analyze this product review...',
maxTokens: 500,
requirements: {
minProviderStake: 10000,
maxLatency: 5000,
verification: 'optimistic'
}
});Provider Selection
Network selects providers based on:
- Stake: Higher stake = more trusted
- Reputation: Historical performance
- Capacity: Available compute
- Latency: Geographic proximity
Execution
Provider runs inference:
- Load model weights (verified hash)
- Execute inference
- Generate proof of computation
- Submit result and proof
Verification
Two verification modes:
Optimistic: Result accepted unless challenged
Result submitted → 24h challenge period → Finalized
↓
Challenge raised
↓
Full re-execution by validators
↓
Slash dishonest partyImmediate: Multiple providers execute, consensus on result
Job → 3 providers execute → 2/3 agreement → Result finalized
↓
Disagreement
↓
Additional validatorsEconomics
Pricing
Market-based pricing with:
- Base compute cost (GPU-hours)
- Model licensing (if applicable)
- Verification premium
- Priority fee
Example: LLaMA-70B, 500 tokens
- Compute: $0.002
- Verification: $0.0005
- Total: $0.0025
vs. Centralized API: $0.004
Savings: 37%Provider Economics
Providers earn:
- Inference fees (80% of payment)
- Staking rewards (from protocol inflation)
- Slashing protection (stake at risk)
Token Utility
ACI token used for:
- Payment for inference
- Provider staking
- Governance voting
- Fee burning (deflationary)
Technical Architecture
Consensus Layer
Built on Lux Network for:
- Sub-second finality
- High throughput (10,000+ TPS)
- Low fees ($0.001 per transaction)
- Post-quantum security
Model Registry
On-chain registry of:
- Model metadata
- Weight hashes
- License terms
- Version history
contract ModelRegistry {
struct Model {
string name;
bytes32 weightsHash;
address owner;
uint256 licenseFeeBps;
bool openAccess;
}
mapping(bytes32 => Model) public models;
function register(Model calldata model) external;
function verify(bytes32 modelId, bytes32 weightsHash) external view returns (bool);
}Proof of Inference
Cryptographic proof that inference ran correctly:
- Input commitment
- Model hash verification
- Output determinism check
- Hardware attestation
Use Cases
Redundant AI Infrastructure
Never depend on single provider:
const result = await aci.inference({
model: 'llama-70b',
prompt: prompt,
fallback: {
providers: 3,
consensus: 'any'
}
});Verifiable AI
Prove AI was used correctly:
const { result, proof } = await aci.inference({
model: 'llama-70b',
prompt: prompt,
verification: 'immediate'
});
// Proof verifiable on-chain
const verified = await aci.verifyProof(proof);Decentralized Applications
AI for dApps without centralized dependencies:
contract AIOracle {
function requestInference(
bytes32 modelId,
bytes calldata prompt
) external returns (bytes32 requestId);
function fulfillInference(
bytes32 requestId,
bytes calldata result,
bytes calldata proof
) external;
}Roadmap
- Q4 2024: Testnet launch
- Q1 2025: Mainnet beta
- Q2 2025: Full mainnet
- Q3 2025: Enterprise features
Get Involved
- Providers: Run nodes, earn rewards
- Developers: Build on ACI
- Token holders: Participate in governance
Documentation: aci.hanzo.ai
The future of AI is decentralized. ACI makes it real.
Zach Kelling is the founder of Hanzo Industries.
Read more
Building a Blockchain Faucet That Doesn't Get Drained
How we built the Lux testnet faucet to resist abuse while remaining accessible to developers.
Netrunner: Testing Blockchains at Scale
How we built Netrunner to simulate blockchain networks and catch bugs before they hit production.
arcalabs: Deploying Geth at Scale for Commerce
Why Hanzo needed its own Geth deployment tooling in 2017, what the arcalabs/geth-node project does, and how Ethereum integration with commerce infrastructure evolved into broader blockchain data tooling.