
Vendor » Cerebras Systems
Features
Cerebras Hits the Accelerator for Deep Learning Workloads
When it comes to large neural networks like BERT or GPT-3, organizations often must wait weeks or even months for a training task to complete if they’re using traditional CPU and GPU clusters. But with its massive Wafe Read more…
A Wave of Purpose-Built AI Hardware Is Building
Google last week unveiled the third version of its Tensor Processing Unit (TPU), which is designed to accelerate deep learning workloads developed in its TensorFlow environment. But that's just the start of a groundswell Read more…
This Just In
ABU DHABI, United Arab Emirates and SUNNYVALE, Calif., Aug. 28, 2025 — Cerebras and Core42 have announced the global availability of OpenAI’s gpt-oss-120B. Core42 AI Cloud via Compass API brings Cerebras Inference at 3,000 tokens per second to power enterprise-scale agentic AI. Read more…
PARIS, July 9, 2025 — At the RAISE Summit in Paris, France, Cerebras Systems announced that Cerebras Inference Cloud is now available in AWS Marketplace bringing Cerebras’ ultra-fast AI inference to enterprise customers, and enabling the next era of high performance, interactive, and intelligent agentic AI applications. Read more…
PARIS and SUNNYVALE, Calif., July 9, 2025 — Cerebras Systems has announced that Notion, the all-in-one connected workspace, is using Cerebras’ industry-leading AI inference technology to power instant, enterprise-scale document search for its AI offering, Notion AI for Work. Read more…
PARIS, July 8, 2025 — Today at the RAISE Summit in Paris, France, Cerebras Systems announced new partnerships and integrations with Hugging Face, DataRobot and Docker. These collaborations dramatically increase accessibility and impact of Cerebras’ ultra-fast AI inference, enabling a new generation of performant, interactive, and intelligent agentic AI applications. Read more…
PARIS, July 8, 2025 — Cerebras Systems today announced the launch of Qwen3-235B with full 131K context support on its inference cloud platform. Alibaba’s Qwen3-235B delivers model intelligence that rivals frontier models such as Claude 4 Sonnet, Gemini 2.5 Flash, and DeepSeek R1 across a range of science, coding, and general knowledge benchmarks, according to independent tests by Artificial Analysis. Read more…
May 9, 2025 — Editor’s Note: IBM and Cerebras Systems have announced a collaboration to integrate Cerebras’ AI computing hardware with IBM’s watsonx platform. The goal is to help enterprises run generative AI models at scale while meeting requirements for trust, efficiency, and integration in complex environments. Read more…
SUNNYVALE, Calif., March 11, 2025 — Cerebras Systems today announced the launch of six new AI inference datacenters powered by Cerebras Wafer-Scale Engines. These state-of-the-art facilities, equipped with thousands of Cerebras CS-3 systems, are expected to serve over 40 million Llama 70B tokens per second, making Cerebras the world’s #1 provider of high-speed inference and the largest domestic high speed inference cloud. Read more…
SUNNYVALE, Calif., March 11, 2025 — Cerebras and Hugging Face today announced a new partnership to bring Cerebras Inference to the Hugging Face platform. HuggingFace has integrated Cerebras into HuggingFace Hub, bringing the world’s fastest inference to over five million developers on HuggingFace. Read more…
SUNNYVALE, Calif., Feb. 13, 2025 — Cerebras Systems has announced its pivotal role in powering Sonar, an advanced model optimized for Perplexity search. Read more…