Cerebras logo

Cerebras

Operational

Wafer-scale AI inference at record speeds

Ultra-fast inference powered by the world's largest AI chip.

uptime

Company

Headquarters
Sunnyvale, CA
Founded
2016

Capabilities

Models Hosted
Curated Deck
Inference Speed
2000+ tokens/sec
Specialties
Ultra-high throughput
Unique Features
Wafer-scale chips
API Style
OpenAI-compatible
Compute Location
US + Canada + EU

Models

Coming soon

We are standardizing model listings across providers.

Why Use Cerebras

Record Speed

2000+ tokens/second on wafer-scale chips.

Free Tier

Generous free tier for developers.

Details

About Cerebras

Cerebras built the world’s largest chip for unprecedented inference speed.

Newsletter

Get the signal, skip the noise.

Weekly digest of new models and provider updates across 41+ compute providers. Curated for AI builders who ship.

New model releases
Capability updates
Provider status
bots.so
The AI Inference Model Index
© bots.so — The AI Inference Model Index

bots.so aggregates publicly available model deployment information from official provider sources. We are not affiliated with any model provider. Model availability changes rapidly; always verify on official sites.