The Chinese trained an analogue of GPT-4 with only 2000 chips and 33 times cheaper than OpenAI

Image source: Copilot

01.ai’s achievement is especially noteworthy given the limited access Chinese companies have to Nvidia’s advanced GPUs. Founder and CEO Kai-Fu Lee points out that despite Chinese companies having virtually no access to Nvidia GPUs due to US regulations, Yi-Lightning’s AI model ranked sixth in performance rankings models according to the LMSIS version of the University of California at Berkeley.

Image source: NVIDIA

«My friends in Silicon Valley are shocked not only by our performance, but also by the fact that we trained a model for only $3 million,” said Kai-Fu Lee. “It is rumored that approximately $1 billion has already been invested in GPT-5 training.” He also added that due to US sanctions, companies in China are forced to look for more efficient and cost-effective solutions, which is what 01.ai was able to achieve by optimizing resources and engineering ideas, while obtaining similar results to GPT-4 at significantly lower costs.

Instead of increasing computing power as competitors are doing, the company has focused on optimizing algorithms and reducing processing bottlenecks. “When we only have 2,000 GPUs, we have to figure out how to use them [effectively],” Lee said.

As a result, model output costs were only 10 cents per million tokens, which is approximately 30 times less than similar models. “We turned the computational problem into a memory problem by building a multi-level cache, creating a special inference engine, and so on,” Li shared the details.

Despite claims about the low cost of training the Yi-Lightning model, questions remain regarding the type and number of GPUs used. The head of 01.ai claims that the company has enough resources to implement its plans for a year and a half, but a simple calculation shows that 2,000 modern Nvidia H100 GPUs at the current price of $30,000 per unit would cost $6 million, which is double the stated costs. This discrepancy raises questions and requires further clarification. However, the company’s achievement has already attracted the attention of the world community and showed that innovation in the field of AI can be born even in conditions of limited computing resources.

admin

Share
Published by
admin

Recent Posts

Microsoft has closed the Windows 10 beta channel again

Microsoft has closed the Windows 10 beta channel and moved all registered Insider users to…

47 minutes ago

Nintendo and Sony are facing a disastrous holiday season – the consoles are old, there are no major releases

The list of offers from Japanese gaming giants Nintendo and Sony this holiday season is…

57 minutes ago

Smart ring maker Oura aims to conquer the international market

Finnish Oura, the world's largest manufacturer of smart rings, did not see a threat to…

2 hours ago

“He doesn’t know how to make a battery”: the head of CATL promises failure for Elon Musk’s 4680 cells

Introduced in 2020 and developed with the participation of Panasonic specialists, the 4680 battery cells…

2 hours ago