AI power consumption was reduced by 95% without loss, but Nvidia is unlikely to approve the new algorithm

With the growing popularity of artificial intelligence, high power consumption of AI models is becoming an increasingly pressing problem. Despite the fact that tech giants such as Nvidia, Microsoft and OpenAI have not yet spoken loudly about this problem, clearly downplaying its significance, specialists from BitEnergy AI have developed a technology that can significantly reduce energy consumption without significant losses in the quality and speed of AI work.

Image source: Copilot

According to the study, the new method can reduce energy use by up to 95%. The team calls their discovery Linear-Complexity Multiplication, or L-Mul for short. According to TechSpot, this computing process is based on adding integers and requires significantly less energy and operations compared to floating-point multiplication, which is widely used in AI-related tasks.

Today, floating point numbers are actively used in AI to process very large or very small numbers. They resemble notation in binary form, allowing algorithms to accurately perform complex calculations. However, such accuracy requires extremely large resources and already raises certain concerns, since some AI models require huge amounts of electricity. For example, ChatGPT requires as much electricity as 18,000 US households consume—564 TWh daily. Analysts at the Cambridge Center for Alternative Finance estimate that by 2027, the AI ​​industry could consume between 85 and 134 TWh annually.

The L-Mul algorithm solves this problem by replacing complex floating-point multiplication operations with simpler integer additions. During testing, AI models maintained accuracy, while energy consumption for tensor operations was reduced by 95%, and for scalar operations by 80%.

L-Mul also improves performance. The algorithm was found to outperform current 8-bit precision computing standards, providing higher precision with fewer bit-level operations. In tests covering a variety of AI tasks, including natural language processing and computer vision, the performance hit was only 0.07%, which experts considered a minor loss compared to the huge energy savings.

That being said, transformer-based models such as GPT may benefit the most from L-Mul as the algorithm is easily integrated into all key components of these systems. And tests on popular AI models such as Llama and Mistral even showed improvements in accuracy in some tasks.

The bad news is that L-Mul requires specialized hardware and current AI accelerators are not optimized to use this method. The good news is that work is already underway to create such hardware and application programming interfaces (APIs).

One possible obstacle could be resistance from large chip makers like Nvidia, which could slow the adoption of the new technology. Since, for example, Nvidia is a leader in the production of equipment for artificial intelligence and it is unlikely that it will so easily give way to more energy-efficient solutions.

admin

Share
Published by
admin

Recent Posts

Express test of external SSD-drive MSI Datamag 20Gbps

Today we will talk about a new gadget from MSI, which the manufacturer itself mysteriously…

2 hours ago

Apple to Release Updated MacBook Air with M4 Chip in March 2025

Apple is preparing to launch updated 13- and 15-inch versions of the MacBook Air laptop,…

3 hours ago

Official Radeon RX 9070 XT Relative Performance Leaked to Press

The VideoCardz portal writes that AMD held a closed briefing for journalists this week, where…

4 hours ago

Kindergarten of some kind: former German data center converted into preschool

Bonn, Germany, is in dire need of kindergartens, so they are sometimes placed in the…

4 hours ago

Apple to Improve iPhone 17 Pro Camera with Focus on Video

According to online sources, Apple will focus more on improving video recording in the new…

4 hours ago

GeForce RTX 5070 Ti with “fallen off” ROPs loses up to 11% performance in synthetic tests

It was previously reported that some GeForce RTX 5090/RTX 5090D graphics cards, and as it…

4 hours ago