List of Cerebras Cloud Customers
Sunnyvale, 94085, CA,
United States
Since 2010, our global team of researchers has been studying Cerebras Cloud customers around the world, aggregating massive amounts of data points that form the basis of our forecast assumptions and perhaps the rise and fall of certain vendors and their products on a quarterly basis.
Each quarter our research team identifies companies that have purchased Cerebras Cloud for AI infrastructure from public (Press Releases, Customer References, Testimonials, Case Studies and Success Stories) and proprietary sources, including the customer size, industry, location, implementation status, partner involvement, LOB Key Stakeholders and related IT decision-makers contact details.
Companies using Cerebras Cloud for AI infrastructure include: Mistral AI, a France based Professional Services organisation with 140 employees and revenues of $65.0 million, Hugging Face, a United States based Professional Services organisation with 500 employees and revenues of $50.0 million, Perplexity, a United States based Professional Services organisation with 55 employees and revenues of $5.0 million and many others.
Contact us if you need a completed and verified list of companies using Cerebras Cloud, including the breakdown by industry (21 Verticals), Geography (Region, Country, State, City), Company Size (Revenue, Employees, Asset) and related IT Decision Makers, Key Stakeholders, business and technology executives responsible for the software purchases.
The Cerebras Cloud customer wins are being incorporated in our Enterprise Applications Buyer Insight and Technographics Customer Database which has over 100 data fields that detail company usage of software systems and their digital transformation initiatives. Apps Run The World wants to become your No. 1 technographic data source!
Apply Filters For Customers
| Logo | Customer | Industry | Empl. | Revenue | Country | Vendor | Application | Category | When | SI | Insight |
|---|---|---|---|---|---|---|---|---|---|---|---|
|
|
Hugging Face | Professional Services | 500 | $50M | United States | Cerebras Systems | Cerebras Cloud | AI infrastructure | 2025 | n/a |
In 2025, Hugging Face integrated Cerebras Cloud into the Hugging Face Hub and Inference API to provide CS-3 powered inference for popular open-source models. Cerebras Cloud is an AI infrastructure provider that was added as an available inference provider on the developer platform, supporting a developer-platform inference use case while Hugging Face is headquartered in the United States and serves a global developer audience.
The implementation exposed explicit inference and integration modules within the Hugging Face Inference API, enabling low-latency, high-throughput model serving for open-source models. Configuration centered on provider selection and endpoint provisioning so developers could route inference requests to Cerebras Cloud CS-3 resources through the Hub and API integration.
Integration scope covered the Hugging Face Hub and the Inference API as the primary operational touchpoints, with Cerebras Cloud providing CS-3 hardware-accelerated inference capacity. The deployment addressed developer platform and model serving business functions, preserving Hub workflows while adding Cerebras as a selectable inference provider for model hosting and runtime inference.
Governance and rollout focused on making Cerebras Cloud available as an inference provider within the existing developer experience, with the intent of expanding developer access to high-throughput model serving. The announced outcome was expanded developer access to CS-3 powered, low-latency inference through Cerebras Cloud as an available inference provider on the Hugging Face Hub and Inference API.
|
|
|
Mistral AI | Professional Services | 140 | $65M | France | Cerebras Systems | Cerebras Cloud | AI infrastructure | 2025 | n/a |
In 2025, Mistral AI deployed Cerebras Cloud to power its Le Chat conversational assistant. The deployment used Cerebras Systems' cloud offering as core AI infrastructure for inference focused model serving in Europe.
The implementation prioritized low latency and extreme throughput, instrumenting Cerebras Cloud compute to run production large language model inference workflows. Functional capabilities implemented included streaming token generation, throughput optimized batching and model serving pipelines, and conversational session handling for Le Chat. Configuration emphasized inference scale and latency controls rather than training workloads.
Mistral reported that the Cerebras Cloud deployment achieved roughly 1,000 words per second in inference and materially reduced end user response latency, delivering record breaking inference throughput for its conversational assistant. Operational scope was inference focused, supporting customer facing chat interactions across Mistral AI’s European deployment footprint. Governance centered on production inference orchestration and latency monitoring.
|
|
|
Perplexity | Professional Services | 55 | $5M | United States | Cerebras Systems | Cerebras Cloud | AI infrastructure | 2025 | n/a |
In 2025, Perplexity deployed Cerebras Cloud to power its Sonar search model built on Llama 3.3 70B, using the vendor platform as the core AI infrastructure for inference. The Cerebras Cloud implementation targeted faster, more factual search answers for Perplexity Pro users and was framed as a search and AI inference deployment.
The deployment provisioned dedicated model hosting and low latency inference capacity, with Cerebras Cloud serving the Sonar model to support real-time query handling. Functional capabilities implemented included low-latency model serving, request routing and concurrency controls, and observability for inference performance to monitor throughput and latency during peak search activity.
This search and AI inference implementation was scoped to North America and focused on improving Perplexity Pro customer experiences, with operational ownership sitting with product and engineering teams. The integration of Cerebras Cloud with the Sonar model produced markedly improved inference speed and efficiency, enabling real-time search experiences for customers.
|
Buyer Intent: Companies Evaluating Cerebras Cloud
- Ace Data Centers, a United States based Professional Services organization with 15 Employees
Discover Software Buyers actively Evaluating Enterprise Applications
| Logo | Company | Industry | Employees | Revenue | Country | Evaluated | ||
|---|---|---|---|---|---|---|---|---|
| No data found | ||||||||