Deepseek x Cerebras: How the most controversial AI model right now is supercharged by the most powerful AI Superchip ever built


  • Maker of the fastest AI -chip in the world makes a splash with deepseek onboarding
  • Cerebras says the solution will rank 57x faster than on GPUs but don’t mention which GPUs
  • Deepseek R1 is running on Cerebras Cloud and the data stays in the US

Cerebras has announced that it will support Deepseek in a non-so-surprising feature, more specifically the R1 70B reasoning model. The move comes after Groq and Microsoft confirmed that they would also bring the new child of the AI ​​block to their respective clouds. AWS and Google Cloud have yet to do so, but anyone can run the Open Source model everywhere, even locally.

AI Inference Chip Specialist will run Deepseek R1 70B at 1,600 tokens/second as it claims is 57x faster than any R1 provider using GPUs; One can deduce that 28 tokens/second is what the GPU-in-the-Cloud solution (in this case Deepin from) apparently reaches. Serendipithed is Cerebra’s latest chip 57x larger than H100. I have reached Cerebras to find out more about this claim.

Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top