Cerebras Systems launched the “world’s most powerful” AI platform for inference

The American startup Cerebras Systems, which develops chips for machine learning systems and other resource-intensive tasks, announced the launch of what is claimed to be the world’s most productive AI platform for inference – Cerebras Inference. It is expected that it will seriously compete with solutions based on NVIDIA accelerators.

The Cerebras Inference cloud system is based on WSE-3 accelerators. These giant products, made using TSMC’s 5nm process technology, contain 4 trillion transistors, 900 thousand cores and 44 GB of SRAM. The total bandwidth of the built-in memory reaches 21 PB/s, and the internal interconnect – 214 PB/s. For comparison, a single HBM3e chip in the NVIDIA H200 boasts a throughput of “only” 4.8 TB/s.

Image source: Cerebras

According to Cerebras, the new inference platform provides up to 20 times higher performance compared to comparable solutions on NVIDIA chips in hyperscaler services. In particular, the performance is up to 1800 tokens per second per user for the Llama3.1 8B AI model and up to 450 tokens per second for Llama3.1 70B. For comparison, for AWS these values ​​are 93 and 50, respectively. We are talking about FP16 operations. Cerebras states that the best result for NVIDIA H100 based clusters in the case of Llama3.1 70B is 128 tokens per second.

«Unlike alternative approaches that sacrifice accuracy for speed, Cerebras offers the highest performance while maintaining 16-bit accuracy for the entire inference process,” the company says.

At the same time, Cerebras Inference services cost several times less compared to competing offers: $0.1 per 1 million tokens for Llama 3.1 8B and $0.6 per 1 million tokens for Llama 3.1 70B. Pay as you go. Cerebras plans to provide inference services through an API compatible with OpenAI. The benefit of this approach is that developers who have already built applications based on GPT-4, Claude, Mistral or other cloud AI models will not have to completely change their code to migrate workloads to the Cerebras Inference platform.

For larger businesses, the Enterprise Tier service plan offers highly customized models, customized experiences and dedicated support. The standard Developer Tier package requires a subscription price starting from $0.1 per 1 million tokens. In addition, there is a free entry-level Free Tier access with restrictions. Cerebras says the launch of the platform will open up entirely new opportunities for the implementation of generative AI in various fields.

admin

Share
Published by
admin

Recent Posts

Nissan Leaf EV to Become NACS-Ported Compact Crossover in Third Generation

Nissan Leaf can rightfully be considered a long-liver of the electric car market, since the…

5 days ago

OpenAI expects to more than triple its revenue this year and then double it next year.

OpenAI, the market leader in generative artificial intelligence systems, remains nominally a startup, its financial…

5 days ago

OpenAI Decides to Hold 4o Image Generation Launch for Free Users

OpenAI has been forced to delay the release of ChatGPT's built-in image generator for free…

5 days ago

1440p and 240Hz for just $200: Xiaomi updates the 27-inch Redmi G27Q gaming monitor

Xiaomi continues to update its Redmi G27Q gaming monitor every year. The model was first…

6 days ago

Beware, Android is shutting down: OS development will cease to be public, but there is no reason to panic

Android device makers can significantly customize the look and feel of the operating system, but…

6 days ago

Fake GeForce RTX 4090s with RTX 3090 chips have started popping up in China — craftsmen are even changing the GPU markings

In China, scammers have started selling GeForce RTX 3090 graphics cards, passing them off as…

6 days ago