Nvidia CEO Jensen Huang said the performance of Nvidia chips has increased 1,000-fold over the past 10 years, defying Moore’s Law. And that progress isn’t slowing down anytime soon, with the company’s chips driving further cost reductions and accelerating artificial intelligence (AI) development, Huang told TechCrunch after his CES 2025 keynote.

Image source: NVIDIA

Moore’s Law, formulated in 1965 by Intel co-founder Gordon Moore, predicted that the number of transistors on a chip would double every two years, resulting in a doubling of chip performance. The cost of computing decreased accordingly. For decades, this rule determined the development of computing technology. However, in recent years, Moore’s Law has begun to slow down. However, Jensen Huang strongly disagrees with this, as he has repeatedly stated before. In November, Huang said the AI ​​world was on the verge of a “hyper Moore’s Law.”

This time, the founder of Nvidia noted that Nvidia’s AI chips are developing ahead of schedule and the new superchip for data centers is 30 times faster than the previous generation of chips when performing tasks related to artificial intelligence. “Our systems are progressing much faster than Moore’s Law,” Huang told TechCrunch on Tuesday.

Huang’s announcement comes as the industry faces questions about slowing progress in artificial intelligence. But Nvidia, while remaining a key player in the market that supplies chips to leading AI labs such as Google, OpenAI and Anthropic, says it can move faster than Moore’s Law because innovation is happening at every level, from chip architecture to software algorithms. “We can create architecture, chip, system, libraries and algorithms all at the same time,” Huang noted. “If you do that, you can move faster than Moore’s Law because you can innovate across the entire stack.”

The Nvidia chief showed off the data center superchip at CES that powers the GB200 NVL72 system, which he says is 30 to 40 times faster at AI computing than the previous flagship H100 chip. This significant increase in performance, according to Huang, will reduce the cost of running AI models that require large computing power, such as OpenAI’s o3 model. It also emphasizes that in the long term, expensive reasoning models can be used to generate better data for subsequent training of AI agents, which will lead to lower costs.

Huang rejects the idea that AI progress is slowing down and argues that advances in hardware could directly impact the future development of AI capabilities. “Moore’s Law was so important in the history of computing because it brought down the cost of computing,” Huang told TechCrunch. “The same thing will happen with inference [running trained neural networks]: we will increase productivity, and as a result, the cost of inference will become less.”

admin

Share
Published by
admin

Recent Posts

Windows Remote Access Has a ‘Hole’ Microsoft Won’t Fix

Researchers have once again drawn attention to a serious vulnerability in the Windows Remote Desktop…

8 minutes ago

Windows Remote Access Has a ‘Hole’ Microsoft Won’t Fix

Researchers have once again drawn attention to a serious vulnerability in the Windows Remote Desktop…

13 minutes ago

Android 16 could get one of the most anticipated features – Google is testing a desktop mode with windows and more

Android 16 will feature a built-in Desktop Mode, which will turn your smartphone into a…

13 minutes ago

Skype to Shut Down Permanently on May 5, Microsoft to Move Users to Teams

On May 5, 2025, Skype will officially cease operations. Microsoft announced the closure of one…

3 hours ago