The Qualcomm® Cloud AI 100 is a performance and cost optimized AI inference solution,
purpose designed for Generative AI, Large Language Models, Natural Language Processing,
and Computer Vision.

Purpose-built for high performance, low-power AI processing in the cloud.

The Qualcomm Cloud AI 100, designed for AI inference acceleration, addresses unique requirements in the cloud, including power efficiency, scale, process node advancements, and signal processing—facilitating the ability to run inference faster and more efficiently. Qualcomm Cloud AI 100 is designed to be a leading solution for customers who are scaling AI inference workloads at their enterprise, across the globe.

Qualcomm’s industry-leading solutions utilize over a decade of research and development delivering high-performance, low-power deep learning inference acceleration technology. This scalable architecture enables AI processing and analytics in both real-time and offline multi-media streams.

Ready to Go?

Sign-up to access the Qualcomm Cloud AI 100 and run inference on the edge cloud faster and more efficiently.


Qualcomm Cloud AI 100 and the Cirrascale AI Innovation Cloud

We've partnered with Qualcomm to offer their cutting-edge AI inference accelerator in the cloud for customers to test, utilize and fully deploy in the cloud. No matter your application, whether large language models (LLM), natural language processing (NLP), or Object Detection, the Cirrascale AI Innovation Cloud utilizing the Qualcomm Cloud AI 100 is for you.

Our flat rate, no surprises billing model means we can provide you with a price for Qualcomm Cloud AI 100 instances that won't fluctuate so you can count on what we've presented as your final price. We also don't nickel-and-dime you by charging to get your data in to or out of our cloud. Instead, we charge no ingress or egress fees, so you never receive a supplemental bill.

Cloud AI 100

Use Cases

The Qualcomm Cloud AI 100 accelerator enables high performance deep learning inference across computer vision, object detection, natural language processing, generative AI models, and more.

Typical use cases of LLMs include: Text-to-code, greatly accelerated application development and site building, customer service and chatbots for retailer online shopping, document summarization and copilot-like usage to summarize meetings or emails, language translation, and improving business access to markets across geographies.

The Qualcomm Cloud AI 100 supports dozens of NLP models like GPT2 and its variants, and Bidirectional Encoder Representations from Transformers (BERT) and its variants. Beyond NLP, Qualcomm Cloud AI 100 supports models in domains from computer vision (image classification, object detection, semantic segmentation, pose estimation, face detection) to autonomous driving.

Qualcomm Inference Performance

Qualcomm Cloud AI 100 Benchmark Results

These latest results demonstrate Qualcomm Cloud AI 100 leadership across all AI inferencing applications for both the datacenter and edge categories by delivering the highest number of inferences at lowest latency and lowest energy utilization. The Cloud AI 100 provides a unique blend of high computational performance, low latency and low power utilization and is well suited for a broad range of applications.



Qualcomm Cloud AI 100 Pricing

All pricing below is based on Cirrascale's No Surprises billing model. There are no hidden fees and discounts may apply for long-term commitments depending on the service requested. All pricing shown for servers are per server per month.

Config vCPUs System RAM Local Storage Monthly Pricing Annual Pricing
Single AI 100 (48) 12 48GB 1TB NVMe $329 $259
Single AI 100 (64) 32 64GB 1TB NVMe $369 $289
Single AI 100 (128) 32 128GB 1TB NVMe $549 $439
Dual AI 100 24 48GB 1TB NVMe $629 $519
Quad AI 100 48 182GB 1TB NVMe $1,259 $1,009
Octo AI 100 64 384GB 1TB NVMe $2,499 $2,019

Get Started Today!

Sign-up to access the Qualcomm Cloud AI 100 and experience unprecedented AI inferencing.