Industrial manufacturing
Industrial Internet of Things | Industrial materials | Equipment Maintenance and Repair | Industrial programming |
home  MfgRobots >> Industrial manufacturing >  >> Industrial Internet of Things >> Embedded

Qualcomm Unveils Cloud AI 100: A Power‑Efficient Edge AI Accelerator for 5G and Beyond

Nearly 18 months after its initial unveiling in April 2019, Qualcomm has now confirmed that the Cloud AI 100 silicon is in production and will ship during the first half of 2021. The company also disclosed the card form factors and performance benchmarks that underscore its readiness for edge‑AI workloads.

While Qualcomm has long dominated the smartphone market with its Snapdragon line, its earlier server‑grade Centriq line was discontinued in 2018. The Cloud AI 100 marks Qualcomm’s strategic entry into the near‑edge server market, delivering high‑throughput AI inference with exceptional power efficiency.

Qualcomm Unveils Cloud AI 100: A Power‑Efficient Edge AI Accelerator for 5G and Beyond
Qualcomm’s Cloud AI 100 is designed for near‑edge AI workloads such as 5G infrastructure and enterprise data centers.

Form Factors and Performance

Qualcomm will initially offer the Cloud AI 100 in three card formats:

These figures represent the theoretical “raw” TOPS and do not account for real‑world application overhead. Nonetheless, Qualcomm’s own performance curves show the PCIe version outperforming many leading solutions while consuming only a fraction of the power.

Qualcomm Unveils Cloud AI 100: A Power‑Efficient Edge AI Accelerator for 5G and Beyond
ResNet‑50 inference throughput versus power consumption. Qualcomm’s cards lead in power efficiency, with a batch size of 8 (except Nvidia A100).

John Kehrli, Qualcomm’s Senior Director of Product Management, noted, “We bring over a decade of AI R&D from our mobile platforms to this new core. While the architecture differs from Snapdragon, the expertise and performance per watt are evident.”

Chip Architecture

The Cloud AI 100 is an inference accelerator built on a 7 nm FinFET process. It houses up to 16 AI cores that support INT8, INT16, FP16, and FP32 precision. The die contains 144 MB of on‑chip SRAM (128 MB shared, 1 MB per core) and interfaces with up to 32 GB of LPDDR4x DRAM (4×64 LPDDR4x at 2.1 GHz), delivering a high memory bandwidth essential for AI workloads.

Target Applications

The accelerator is tailored for a wide range of edge use cases, including computer vision, speech recognition, autonomous driving, language translation, and recommendation engines. Qualcomm is positioning the Cloud AI 100 for four primary markets:

In the autonomous driving space, a derivative of the Cloud AI 100 powers Qualcomm’s Snapdragon Ride platform, leveraging the same architecture and software stack.

Reference Design & Development Kit

Qualcomm introduced a new development kit featuring a reference design for a 5G edge box powered by the Cloud AI 100. The kit pairs the accelerator with a Snapdragon 865 host processor, offering full video pipelines that support up to 24 streams of full‑HD video decode. It also includes a pre‑certified Snapdragon X55 5G modem.

Kehrli emphasized, “This kit enables customers to launch demo applications immediately. It comes pre‑compiled with ResNet‑50, providing a quick start for edge AI solutions.”

Deployment Timeline

Qualcomm expects initial commercial deployments to focus on edge scenarios—smart cities, retail analytics, and manufacturing—due to shorter certification cycles. Sampling of the Cloud AI 100 is already underway with select customers, and the development kit will sample next month for a limited audience.

Note: This article was originally published on EE Times.

Embedded

  1. Serverless Computing: The Next Generation of “As a Service” Innovation
  2. Edge & Cloud Computing in IoT: A Concise Evolutionary Overview
  3. Edge, Endpoint, and Cloud AI: A Unified Future for Intelligent IoT
  4. Edge AI Chips: Driving the Future of On-Device Intelligence
  5. Qualcomm Unveils Cloud AI 100: A Power‑Efficient Edge AI Accelerator for 5G and Beyond
  6. Advantech Expands Accelerated Computing to the Edge with NVIDIA Integration
  7. ADLINK Deploys AI from Edge to Cloud with NVIDIA EGX Platform: Accelerated Edge Computing Solutions
  8. IXrouter: Seamless Edge‑to‑Cloud Connectivity for Industrial IoT
  9. Edge Cloud Computing: The Essential Backbone for IoT’s Rapid Growth
  10. Edge Computing vs Cloud Computing: Key Differences Explained