The Chinese trained an analogue of GPT-4 with only 2000 chips and 33 times cheaper than OpenAI

Image source: Copilot

01.ai’s achievement is especially noteworthy given the limited access Chinese companies have to Nvidia’s advanced GPUs. Founder and CEO Kai-Fu Lee points out that despite Chinese companies having virtually no access to Nvidia GPUs due to US regulations, Yi-Lightning’s AI model ranked sixth in performance rankings models according to the LMSIS version of the University of California at Berkeley.

Image source: NVIDIA

«My friends in Silicon Valley are shocked not only by our performance, but also by the fact that we trained a model for only $3 million,” said Kai-Fu Lee. “It is rumored that approximately $1 billion has already been invested in GPT-5 training.” He also added that due to US sanctions, companies in China are forced to look for more efficient and cost-effective solutions, which is what 01.ai was able to achieve by optimizing resources and engineering ideas, while obtaining similar results to GPT-4 at significantly lower costs.

Instead of increasing computing power as competitors are doing, the company has focused on optimizing algorithms and reducing processing bottlenecks. “When we only have 2,000 GPUs, we have to figure out how to use them [effectively],” Lee said.

As a result, model output costs were only 10 cents per million tokens, which is approximately 30 times less than similar models. “We turned the computational problem into a memory problem by building a multi-level cache, creating a special inference engine, and so on,” Li shared the details.

Despite claims about the low cost of training the Yi-Lightning model, questions remain regarding the type and number of GPUs used. The head of 01.ai claims that the company has enough resources to implement its plans for a year and a half, but a simple calculation shows that 2,000 modern Nvidia H100 GPUs at the current price of $30,000 per unit would cost $6 million, which is double the stated costs. This discrepancy raises questions and requires further clarification. However, the company’s achievement has already attracted the attention of the world community and showed that innovation in the field of AI can be born even in conditions of limited computing resources.

admin

Share
Published by
admin

Recent Posts

The US government considers GlobalFoundries a good candidate to save Intel

Until now, it was believed that large suppliers of semiconductor products such as Qualcomm and…

5 minutes ago

Microsoft and Ubisoft have solved the problem of Assassin’s Creed compatibility with Windows 11 24H2

Microsoft has lifted restrictions on updating Windows 11 to version 24H2 for computers running Assassin's…

5 minutes ago

Windows 11 will become smarter: Microsoft is testing AI file search

Microsoft is testing a new artificial intelligence (AI)-powered search feature in the latest build for…

1 hour ago

Merger instead of sale: Perplexity AI wants to save TikTok in the US

Perplexity AI proposed on Saturday, a day before TikTok was blocked in the United States,…

1 hour ago

Battle Shapers – fear of ambition. Review

Not defined Roguelikes with a first-person perspective are a fairly niche genre segment, but they…

6 hours ago

ASRock introduced industrial mini-PCs and motherboards based on Intel Arrow Lake-H and AMD Ryzen 300 AI

ASRock Industrial, according to the CNX-Software resource, presented industrial computers of a small form factor…

7 hours ago