The Qualcomm® Cloud AI 100 is a performance and cost optimized AI inference solution,
purpose designed for Generative AI, Large Language Models, Natural Language Processing,
and Computer Vision.
The Qualcomm Cloud AI 100, designed for AI inference acceleration, addresses unique requirements in the cloud, including power efficiency, scale, process node advancements, and signal processing—facilitating the ability to run inference faster and more efficiently. Qualcomm Cloud AI 100 is designed to be a leading solution for customers who are scaling AI inference workloads at their enterprise, across the globe.
Qualcomm’s industry-leading solutions utilize over a decade of research and development delivering high-performance, low-power deep learning inference acceleration technology. This scalable architecture enables AI processing and analytics in both real-time and offline multi-media streams.
Sign-up to access the Qualcomm Cloud AI 100 and run inference on the edge cloud faster and more efficiently.
SIGN UP TODAYWe've partnered with Qualcomm to offer their cutting-edge AI inference accelerator in the cloud for customers to test, utilize and fully deploy in the cloud. No matter your application, whether large language models (LLM), natural language processing (NLP), or Object Detection, the Cirrascale AI Innovation Cloud utilizing the Qualcomm Cloud AI 100 is for you.
Our flat rate, no surprises billing model means we can provide you with a price for Qualcomm Cloud AI 100 instances that won't fluctuate so you can count on what we've presented as your final price. We also don't nickel-and-dime you by charging to get your data in to or out of our cloud. Instead, we charge no ingress or egress fees, so you never receive a supplemental bill.
The Qualcomm Cloud AI 100 accelerator enables high performance deep learning inference across computer vision, object detection, natural language processing, generative AI models, and more.
Typical use cases of LLMs include: Text-to-code, greatly accelerated application development and site building, customer service and chatbots for retailer online shopping, document summarization and copilot-like usage to summarize meetings or emails, language translation, and improving business access to markets across geographies.
The Qualcomm Cloud AI 100 supports dozens of NLP models like GPT2 and its variants, and Bidirectional Encoder Representations from Transformers (BERT) and its variants. Beyond NLP, Qualcomm Cloud AI 100 supports models in domains from computer vision (image classification, object detection, semantic segmentation, pose estimation, face detection) to autonomous driving.
These latest results demonstrate Qualcomm Cloud AI 100 leadership across all AI inferencing applications for both the datacenter and edge categories by delivering the highest number of inferences at lowest latency and lowest energy utilization. The Cloud AI 100 provides a unique blend of high computational performance, low latency and low power utilization and is well suited for a broad range of applications.
DISCOVER BENCHMARK / PERFORMANCE RESULTSAll pricing below is based on Cirrascale's No Surprises billing model. There are no hidden fees and discounts may apply for long-term commitments depending on the service requested. All pricing shown for servers are per server per month.
Config | vCPUs | System RAM | Local Storage | Monthly Pricing | Annual Pricing |
---|---|---|---|---|---|
Single AI 100 (48) | 12 | 48GB | 1TB NVMe | $329 | $259 |
Single AI 100 (64) | 32 | 64GB | 1TB NVMe | $369 | $289 |
Single AI 100 (128) | 32 | 128GB | 1TB NVMe | $549 | $439 |
Dual AI 100 | 24 | 48GB | 1TB NVMe | $629 | $519 |
Quad AI 100 | 48 | 182GB | 1TB NVMe | $1,259 | $1,009 |
Octo AI 100 | 64 | 384GB | 1TB NVMe | $2,499 | $2,019 |
Sign-up to access the Qualcomm Cloud AI 100 and experience unprecedented AI inferencing.