On October 27, 2025, Qualcomm Technologies made a significant announcement regarding the launch of its AI inference chip solutions tailored for the data center market. These solutions are the Qualcomm AI200 and AI250, which come in both accelerator card and rack system configurations.
The AI200 is meticulously crafted for rack - level AI inference. It boasts support for 768GB of LPDDR memory. This design ensures a low total cost of ownership (TCO) while delivering optimized performance for large language models and multimodal model inference. In the context of the data center industry, where cost - effectiveness and efficient model processing are crucial, the AI200 stands out as a valuable option.
The AI250, on the other hand, embraces a near - memory computing architecture. This innovative architecture brings about a remarkable over 10 - fold improvement in memory bandwidth. It also leads to a substantial reduction in power consumption, which is a key consideration in data centers aiming for energy efficiency. Moreover, it supports distributed inference, a capability that is becoming increasingly important as data processing needs grow more complex and widespread.
Both the AI200 and AI250 are equipped with a range of advanced features. They incorporate direct liquid cooling, which is an efficient way to manage the heat generated by high - performance computing components. They also offer PCIe/Ethernet expansion options, enabling seamless integration with existing data center infrastructures. Confidential computing capabilities are included to ensure data security, a top priority in today's digital landscape. Additionally, the rack - level power consumption is set at 160kW, providing a clear benchmark for power management in data centers.
Accompanying these hardware solutions is a hyperscale AI software stack. This comprehensive software suite covers the entire pipeline of AI operations. It supports mainstream frameworks, making it compatible with a wide range of AI development tools. Furthermore, it allows for one - click deployment of Hugging Face models, simplifying the process of implementing advanced AI models in data centers.
Looking ahead, the AI200 is expected to hit the commercial market in 2026, while the AI250 will follow in 2027.
