Key Points
- Superpods shift the focus from single‑card FLOPS to system‑level efficiency, combining thousands of GPUs with high‑speed optical interconnects to improve end‑to‑end training and inference throughput.
- Alibaba Cloud — PanJiu 128 (磐久128): integrates the in‑house CIPU 2.0 and supports 128 AI compute chips per cabinet, claiming about a 50% inference performance gain versus comparable traditional architectures at the same scale.
- Huawei (Huáwéi 华为) scale and roadmap: sold over 300 CloudMatrix 384 systems serving > 20 customers; describes chaining 432 superpods to reach clusters of ~160,000 cards and product plans like the Atlas 950 (8,192‑card) and Atlas 960 (15,488‑card).
- Operational challenges: many superpod cabinets exceed 100 kW per cabinet, making liquid cooling and improved optical module reliability essential to unlock higher sustained interconnect bandwidth and compute rates.

Superpods are reshaping how China builds AI infrastructure at scale.
Why AI Is Driving a Shift from Single‑Card Performance to Superpods
Artificial intelligence is remapping industries at unprecedented speed.
Model sizes have grown from millions to trillions of parameters, and that drives new infrastructure needs.
Superpods (also written as SuperPoD or 超节点) are replacing single servers and conventional clusters for large‑scale AI workloads.
The term superpod was popularized by NVIDIA (Yīngwěidá 英伟达).
A superpod bundles thousands of GPUs into a single logical unit — effectively a “super compute node.”
These systems rely on high‑speed optical and advanced interconnect technologies to overcome bandwidth and latency bottlenecks between cabinets.
The outcome is higher system‑level efficiency for both training and inference.

Resume Captain
Your AI Career Toolkit:
- AI Resume Optimization
- Custom Cover Letters
- LinkedIn Profile Boost
- Interview Question Prep
- Salary Negotiation Agent

What Alibaba and Huawei Are Shipping
At the 2025 Yunqi (Cloud) Conference, Alibaba Cloud (Ālǐyún 阿里云) unveiled the PanJiu 128 (磐久128) superpod AI server.
Alibaba says the PanJiu 128 integrates the company’s in‑house CIPU 2.0 chip and high‑performance EIC/MOC network cards and supports 128 AI compute chips per cabinet.
Alibaba claims roughly a 50% inference performance gain versus comparable traditional architectures at the same compute scale.
Huawei (Huáwéi 华为) has also doubled down on superpods.
Its CloudMatrix 384 superpod — designed for large training clusters — can be scaled into massive pools by cascading cabinets.
For trillion‑parameter and above models, Huawei describes configurations that link hundreds of superpods.
For example, chaining 432 superpods to form clusters of up to ~160,000 cards for extreme‑scale training in cloud data centers.

Product Roadmaps and Sales Progress
At Huawei’s Full‑Connection Conference in September 2025, the company reported it has sold more than 300 CloudMatrix 384 superpod systems.
Those systems serve over 20 customers, mainly government and enterprise.
Huawei said it will launch the Atlas 950 SuperPoD with an 8,192‑card scale, targeted for Q4 2026.
A follow‑on Atlas 960 SuperPoD, with an expected 15,488‑card scale, is slated for Q4 2027.

Find Top Talent on China's Leading Networks
- Post Across China's Job Sites from $299 / role, or
- Hire Our Recruiting Pros from $799 / role
- Qualified Candidate Bundles
- Lower Hiring Costs by 80%+
- Expert Team Since 2014
Your First Job Post

Other Domestic Players Moving Fast
China’s ecosystem is accelerating beyond the two giants.
-
Inspur (Làngcháo 浪潮 / Inspur) released the “YuanNao SD200” (元脑SD200), positioned for trillion‑parameter models and superpod deployments.
-
MuXi Co., Ltd. (沐曦股份 Mùxī Gǔfèn) has introduced multiple superpod variants, including optical‑interconnect designs, a 3D‑mesh “YaoLong” (耀龙) architecture, Shanghai Cube high‑density liquid‑cooled cabinets, and high‑density liquid‑cooled compute PODs.
-
Baidu (Bǎidù 百度) Intelligent Cloud put its Baige AI computing platform v5.0 into service, formally enabling Kunlun‑chip superpod configurations.

ExpatInvest China
Grow Your RMB in China:
- Invest Your RMB Locally
- Buy & Sell Online in CN¥
- No Lock-In Periods
- English Service & Data
- Start with Only ¥1,000

From Chip Benchmarks to System Efficiency
Securities firms and industry analysts note the competition between the U.S. and China has shifted.
The emphasis is less on single‑card peak FLOPS and more on system‑level efficiency — how quickly, cost‑effectively, and reliably a data center can run large models end‑to‑end.
Domestic players are trying to leapfrog by combining massive cluster builds, open‑source software ecosystems, and industrialized delivery.
Huawei’s superpod approach highlights two advantages.
First, full optical cabinet‑to‑cabinet interconnects that deliver high bandwidth and low latency.
Second, innovations in materials and liquid‑cooling techniques that improve optical module reliability.
Huawei has claimed the Atlas 950’s per‑cabinet card count and aggregate bandwidth are multiple times that of comparable upcoming NVIDIA offerings.
Product comparisons depend heavily on configuration, use case, and the particular metric being measured.

Engineering Challenges: Cooling, Power, and Interconnects
As superpods push compute density up by orders of magnitude, infrastructure challenges follow.
Many superpod cabinets — including Huawei CloudMatrix 384 and NVIDIA GB200/NVL72 class systems — commonly exceed 100 kW per cabinet.
That exponential increase in thermal load and power draw elevates challenges for data‑center cooling, power distribution, and reliability.
Liquid cooling is emerging as the practical solution for the highest densities.
Full‑liquid cooling reduces thermal gradients, enables higher interconnect speeds, and can improve the long‑term reliability of optical modules.
As superpods adopt full‑liquid cooling, both interconnect bandwidth and achievable sustained compute rates are expected to improve materially.

Market and Investment View
Securities analysts note that domestic superpod platforms lead on aggregate metrics like total compute, memory capacity, and interconnect bandwidth for configured systems.
Those attributes could accelerate on‑premises and cloud deployments in China.
Analysts expect rising superpod penetration to drive significant demand across the optical‑interconnect supply chain, liquid‑cooling hardware, and specialized system integrators.
That demand will grow as China builds out AI infrastructure to serve training, inference, and mixed workloads at scale.

What This Means for AI Strategy
For companies building or buying AI infrastructure today, the strategic choice is becoming clearer.
Squeezing marginal single‑card gains matters less than optimizing system throughput, latency, and total cost of ownership at scale.
Superpods emphasize a systems‑level engineering approach — combining hardware, interconnects, cooling, software stack, and deployment logistics.
Many analysts argue this systems approach is the right path for scaling big models in production.
If you’re an investor, founder, or engineer evaluating AI infrastructure, prioritize end‑to‑end systems metrics over chip benchmarks — and track how Superpods change the calculus.
Linking opportunities
-
Alibaba Cloud (Ālǐyún 阿里云) — https://www.alibabacloud.com
-
Huawei (Huáwéi 华为) — https://www.huawei.com
-
Inspur (Làngcháo 浪潮 / Inspur) — https://www.inspur.com
-
Baidu Intelligent Cloud (Bǎidù 百度) — https://cloud.baidu.com
Bottom line: the future of large‑scale AI in China is leaning hard into systems over single‑card peaks, and the industry’s bets on Superpods reflect that shift.

References
- Huawei, Alibaba Bet on “Superpods”: Moving Beyond Single‑Card Performance to System Efficiency — 科创板日报
- PanJiu 128 Superpod Announcement — Alibaba Cloud (阿里云)
- CloudMatrix 384 and Atlas SuperPoD Roadmap — Huawei (华为)
- YuanNao SD200 Superpod — Inspur (浪潮信息)