Inference Provider

VERIFIED
3,888,078 monthly requests

AI & ML interests

None defined yet.

Recent Activity

Cerebras is the inventor of the Wafer-Scale Engine – the revolutionary processor at the heart of our Cerebras CS-X systems.

Our co-designed hardware/software stack enables training large language models upward of 1 trillion parameters using only data parallelism. Our massive on-chip memory enables us to fit all model weights in SRAM and run the World’s Fastest AI Inference.

Browse popular models enabled by Cerebras on Hugging Face!

Explore models trained on Cerebras!

Join the Cerebras Discord to discuss our work and research!