Inference Provider
VERIFIED
3,888,078 monthly requests
AI & ML interests
None defined yet.
Recent Activity
View all activity
Organization Card
Cerebras is the inventor of the Wafer-Scale Engine – the revolutionary processor at the heart of our Cerebras CS-X systems.
Our co-designed hardware/software stack enables training large language models upward of 1 trillion parameters using only data parallelism. Our massive on-chip memory enables us to fit all model weights in SRAM and run the World’s Fastest AI Inference.
Browse popular models enabled by Cerebras on Hugging Face!
Explore models trained on Cerebras!
Join the Cerebras Discord to discuss our work and research!
models
19

cerebras/Llama-3-CBHybridM-8B
Text Generation
•
8B
•
Updated
•
5

cerebras/Llama-3-CBHybridL-8B
Text Generation
•
8B
•
Updated
•
5

cerebras/Dragon-DocChat-Context-Encoder
Updated
•
3
•
2

cerebras/Dragon-DocChat-Query-Encoder
Updated
•
7
•
1

cerebras/Llama3-DocChat-1.0-8B
Text Generation
•
Updated
•
54
•
68

cerebras/Cerebras-GPT-Intermediate
Text Generation
•
Updated

cerebras/Cerebras-LLaVA-13B
Text Generation
•
Updated
•
5
•
5

cerebras/Cerebras-ViT-L-336-patch14-llava13b-ShareGPT4V
Updated
•
2

cerebras/Cerebras-ViT-L-336-patch14-llava7b-ShareGPT4V
Updated
•
4

cerebras/Cerebras-LLaVA-7B
Text Generation
•
Updated
•
6
•
2