Qualcomm introduced energy-efficient AI microservers AI On-Prem Appliance Solution

Qualcomm Technologies announced the Qualcomm AI On-Prem Appliance Solution, a compact, energy-efficient hardware solution for on-premises processing of inference and computer vision workloads. The company also presented a ready-to-use set of AI applications, libraries, models and agents, the Qualcomm Cloud AI Inference Suite, capable of working at the edge in the clouds.

According to the press release, the combination of new products enables SMBs and industrial organizations to run custom and pre-built AI applications on their premises, including generative AI workloads. Qualcomm noted that in-house inference will significantly reduce operating costs and total total cost of ownership (TCO) compared to renting third-party AI infrastructure.

Image Source: Qualcomm

With the AI ​​On-Prem Appliance Solution coupled with the AI ​​Inference Suite, customers will be able to use proprietary data-driven generative AI, fine-tuned models, and technology infrastructure to automate processes and applications in virtually any environment, such as retail stores, restaurants, point-of-sales, dealerships, hospitals, factories and workshops where work processes are well-established, repeatable and ready for automation.

«The AI ​​On-Prem Appliance Solution and Cloud AI Inference Suite change the TCO of AI by enabling generative AI workloads to be processed locally rather than in the cloud,” the company said, highlighting that the AI ​​On-Prem Appliance Solution can significantly reduce application operating costs AI for enterprise and industrial needs in a wide range of areas. In addition, local deployment provides protection against sensitive data leaks.

The Qualcomm AI On-Prem Appliance Solution platform is powered by the Qualcomm Cloud AI family of accelerators. It is reported that the new product supports a wide range of capabilities, including:

  • Scalable from a single desktop device to a wall-mounted cluster;
  • Serving a variety of on-premises AI services, including out-of-the-box voice agents in a box, small language models (SLMs), large language models (LLMs), multimodal models (LMMs), RAGs, agent AI, AI automation platforms, and more;
  • Supports a wide range of generative AI, natural language processing, and computer vision models;
  • Stream AI processing of camera images and video for computer vision applications focused on safety, security and monitoring.

In turn, the Qualcomm Cloud AI Inference Suite offers a complete set of tools and libraries for developing or migrating generative AI applications to the AI ​​On-Prem Appliance Solution or other platforms based on Qualcomm Cloud AI accelerators. The set offers many APIs for user management and administration, for chats, for generating images, audio and video. Claims compatibility with OpenAI API and RAG support. In addition, integration with popular generative AI models and frameworks is available. Deployment using Kubernetes and bare metal is possible.

admin

Share
Published by
admin

Recent Posts

Arm is interested in buying server processor developer Ampere Computing

The British developer of processor architectures Arm is trying to refute periodic rumors about intentions…

7 minutes ago

Microsoft will cut some employees who are not very efficient

In the wake of the artificial intelligence boom, OpenAI's major shareholder, Microsoft, increased its capitalization…

1 hour ago

Thermal Grizzly introduced a device for heating Intel LGA 1851 processors during the process of removing the cover

Until now, all commercial devices for removing covers from Intel processors have facilitated mainly the…

3 hours ago

Micron introduced a new SSD at CES 2025 and showed 64 GB DDR5 modules

Micron has updated its Crucial line by releasing the P510 SSD with PCIe 5.0 interface…

5 hours ago

TSMC began production of AMD Ryzen 9000 and Apple S9 processor components at a plant in the USA

Taiwanese company TSMC has launched production of chips for AMD Ryzen 9000 processors for personal…

5 hours ago