Skip to main content

Products

Solutions

Customers

Developers

Company

James Wang

All
Blog
Press Release

Llama 4 is now available on Cerebras Inference

April 09, 2025

Blog

Cerebras brings instant inference to Mistral Le Chat - Cerebras

February 06, 2025

Blog

Cerebras Launches World's Fastest DeepSeek R1 Llama-70B Inference

January 29, 2025

Blog

100x Defect Tolerance: How Cerebras Solved the Yield Problem - Cerebras

January 13, 2025

Blog

Llama 3.1 405B now runs at 969 tokens/s on Cerebras Inference - Cerebras

November 18, 2024

Blog

Cerebras Inference now 3x faster: Llama3.1-70B breaks 2,100 tokens/s - Cerebras

October 24, 2024

Blog

Introducing Cerebras Inference: AI at Instant Speed - Cerebras

August 27, 2024

Blog

Introducing Sparse Llama: 70% Smaller, 3x Faster, Full Accuracy - Cerebras

May 15, 2024

Blog

Cerebras CS-3 vs. Nvidia B200: 2024 AI Accelerators Compared - Cerebras

April 12, 2024

Blog

Cerebras CS-3: the world’s fastest and most scalable AI accelerator - Cerebras

March 12, 2024

Blog

Cerebras, Petuum, and MBZUAI Announce New Open-Source CrystalCoder and LLM360 Methodology to Accelerate Development of Transparent and Responsible AI Models - Cerebras

December 14, 2023

Press Release

Five Reasons to Join Cerebras in 2024 - Cerebras

December 08, 2023

Blog

Cerebras Software Release 2.0: 50% Faster Training, PyTorch 2.0 Support, Diffusion Transformers, and More

November 10, 2023

Blog

How we fine-tuned Llama2-70B to pass the US Medical License Exam in a week

October 12, 2023

Blog

Introducing Condor Galaxy 1: a 4 exaFLOPS Supercomputer for Generative AI - Cerebras

July 20, 2023

Blog

Introducing Condor Galaxy 1: a 4 exaFLOPS Supercomputer for Generative AI

July 20, 2023

Blog