Accelerator for the accelerator: Fujitsu introduced software that can double the processing speed of AI tasks

Fujitsu announced the availability of middleware designed to optimize the use of AI accelerators. As stated in the press release, this solution makes it possible to increase the efficiency of accelerators, which is especially important in the context of a shortage of computing resources of this type.

The software distinguishes between code that requires a GPU to run and that which can run on the CPU alone, optimizing resource allocation and memory management across different platforms and AI applications. In addition, the software controls the priority of running calculations, giving preference to more efficient processes. Interestingly, allocation does not use the traditional approach of basing resource selection on the entire task.

Image source: Fujitsu

The company announced the solution (Adaptive GPU Allocator) in November 2023. Then they talked about the need to use a proprietary framework based on TensorFlow and PyTorch. The current announcement does not explicitly mention this, but it is reported that the product combines the technology of adaptive resource allocation of each individual accelerator with some kind of AI-based optimization. Moreover, the new product allows you to efficiently process even those tasks that do not fit entirely in the accelerator’s memory. During testing, we were even able to process 150 GB of AI data on a GPU with approximately 30 GB of free RAM.

Fujitsu said the solution increased AI computing efficiency by 2.25 times in real-world testing at AWL, Xtreme-D and Morgenrot. And two large customers, Tradom and Sakura Internet, have already begun to implement the new tool. “By addressing the accelerator and energy shortages caused by the growing global demand for AI, Fujitsu aims to contribute to enhancing business productivity and creativity for its customers,” the company said.

However, so far the solution can only speed up the operation of accelerators in one server, but the company is working to ensure that it can serve multiple GPUs installed in several servers. In other words, it will not yet allow you to speed up an entire AI cluster, but it is still a convenient way to “squeeze more” out of a GPU server, noted The Register resource.

admin

Share
Published by
admin

Recent Posts

Windows 11 will become smarter: Microsoft is testing AI file search

Microsoft is testing a new artificial intelligence (AI)-powered search feature in the latest build for…

50 minutes ago

Merger instead of sale: Perplexity AI wants to save TikTok in the US

Perplexity AI proposed on Saturday, a day before TikTok was blocked in the United States,…

50 minutes ago

Battle Shapers – fear of ambition. Review

Not defined Roguelikes with a first-person perspective are a fairly niche genre segment, but they…

6 hours ago

ASRock introduced industrial mini-PCs and motherboards based on Intel Arrow Lake-H and AMD Ryzen 300 AI

ASRock Industrial, according to the CNX-Software resource, presented industrial computers of a small form factor…

7 hours ago

The potential US Secretary of Transportation promised to deal with SpaceX fines and eliminate the space bureaucracy

This week, Congress held confirmation hearings for new ministers nominated by new US President Donald…

8 hours ago

Vast Space has built the world’s first private space station; it will go into orbit this year

California-based startup Vast Space has announced the completion of the world's first commercial space station,…

8 hours ago