Follow BigDATAwire:

Vendor » Cerebras Systems

Features

OpenXLA Delivers Flexibility for ML Apps

Machine learning developers gained new abilities to develop and run their ML programs on the framework and hardware of their choice thanks to the OpenXLA Project, which today announced the availability of key open source Read more…

Cerebras Hits the Accelerator for Deep Learning Workloads

When it comes to large neural networks like BERT or GPT-3, organizations often must wait weeks or even months for a training task to complete if they’re using traditional CPU and GPU clusters. But with its massive Wafe Read more…

A Wave of Purpose-Built AI Hardware Is Building

Google last week unveiled the third version of its Tensor Processing Unit (TPU), which is designed to accelerate deep learning workloads developed in its TensorFlow environment. But that's just the start of a groundswell Read more…

This Just In

Cerebras Announces 6 New AI Datacenters Across North America and Europe

Mar 11, 2025 |

SUNNYVALE, Calif., March 11, 2025 — Cerebras Systems today announced the launch of six new AI inference datacenters powered by Cerebras Wafer-Scale Engines. These state-of-the-art facilities, equipped with thousands of Cerebras CS-3 systems, are expected to serve over 40 million Llama 70B tokens per second, making Cerebras the world’s #1 provider of high-speed inference and the largest domestic high speed inference cloud. Read more…

Cerebras Partners with Hugging Face to Deliver High-Speed AI Inference

Mar 11, 2025 |

SUNNYVALE, Calif., March 11, 2025 — Cerebras and Hugging Face today announced a new partnership to bring Cerebras Inference to the Hugging Face platform. HuggingFace has integrated Cerebras into HuggingFace Hub, bringing the world’s fastest inference to over five million developers on HuggingFace. Read more…

Cerebras Powers Perplexity Sonar with Industry’s Fastest AI Inference

Feb 12, 2025 |

SUNNYVALE, Calif., Feb. 13, 2025 — Cerebras Systems has announced its pivotal role in powering Sonar, an advanced model optimized for Perplexity search. Read more…

Cerebras Launches Record-Breaking DeepSeek R1 Distill Llama 70B Inference

Jan 30, 2025 |

SUNNYVALE, Calif., Jan. 30, 2025 — Cerebras Systems, a pioneer in accelerating generative AI, today announced record-breaking performance for DeepSeek-R1-Distill-Llama-70B inference, achieving more than 1,500 tokens per second – 57 times faster than GPU-based solutions. Read more…

Cerebras Delivers Record-Breaking Performance with Meta’s Llama 3.1 405B Model

Nov 18, 2024 |

SUNNYVALE, Calif., Nov. 18, 2024 — Cerebras Systems today announced that it has set a new performance record for Llama 3.1 405B – a leading frontier model released by Meta AI. Read more…

Cerebras Systems Announces Filing of Registration Statement for Proposed IPO

Sep 30, 2024 |

SUNNYVALE, Calif., Sept. 30, 2024 — Cerebras Systems today announced that it has filed a registration statement on Form S-1 with the U.S. Securities and Exchange Commission (SEC) relating to a proposed initial public offering of its Class A common stock. Read more…

Cerebras Introduces Inference Service, Touting Speed and Cost Benefits

Aug 27, 2024 |

SUNNYVALE, Calif., Aug. 27, 2024 — Today, Cerebras Systems, a pioneer in high performance AI compute, announced Cerebras Inference. Delivering 1,800 tokens per second for Llama 3.1 8B and 450 tokens per second for Llama 3.1 70B, Cerebras Inference is 20 times faster than NVIDIA GPU-based solutions in hyperscale clouds. Read more…

Cerebras Enables Faster Training of Industry’s Leading Largest AI Models

Jun 13, 2024 |

SUNNYVALE, Calif., June 13, 2024 —Cerebras Systems, a pioneer in accelerating generative artificial intelligence (AI), today announced a collaboration with Dell Technologies, to deliver groundbreaking AI compute infrastructure for generative AI. Read more…

Cerebras and Neural Magic Collaborate to Accelerate Sparse Training and Deployment of LLMs

May 17, 2024 |

SUNNYVALE, Calif. and CAMBRIDGE, Mass., May 17, 2024 — Cerebras Systems, a pioneer in accelerating generative AI, and Neural Magic, a leader in high-performance enterprise inference servers, announced the results of their collaboration for sparse training and deployment of large language models (LLMs). Read more…

Cerebras, Petuum, and MBZUAI Unveil CrystalCoder, a 7 Billion Parameter AI Model

Dec 12, 2023 |

SUNNYVALE, Calif., Dec. 12, 2023 — Cerebras Systems, a pioneer in accelerating generative AI, and Petuum, a generative AI company focused on building transparent LLMs, in partnership with MBZUAI today launched CrystalCoder, a new 7 billion parameter model designed for English language and coding tasks. Read more…

BigDATAwire