Cerebras
OperationalWafer-scale AI inference at record speeds
Ultra-fast inference powered by the world's largest AI chip.
— uptime —
Company
- Headquarters
- Sunnyvale, CA
- Founded
- 2016
Capabilities
- Models Hosted
- Curated Deck
- Inference Speed
- 2000+ tokens/sec
- Specialties
- Ultra-high throughput
- Unique Features
- Wafer-scale chips
- API Style
- OpenAI-compatible
- Compute Location
- US + Canada + EU
Models
Coming soon
We are standardizing model listings across providers.
Why Use Cerebras
Record Speed
2000+ tokens/second on wafer-scale chips.
Free Tier
Generous free tier for developers.
Details
About Cerebras
Cerebras built the world’s largest chip for unprecedented inference speed.
Newsletter
Get the signal, skip the noise.
Weekly digest of new models and provider updates across 41+ compute providers. Curated for AI builders who ship.
New model releases
Capability updates
Provider status