Key Points
- Superpods (超节点, Superpod): a shift from single‑GPU speed to system‑level compute efficiency, treating thousands of accelerators as a single logical unit to improve throughput, latency, and reliability for very large models.
- Alibaba & Huawei product moves: Alibaba Cloud’s Panjiu 128 (磐久128) supports 128 AI compute chips per cabinet and claims inference performance up to 50% higher; Huawei’s CloudMatrix 384 aims at clusters with more than ten thousand accelerators and can cascade 432 CloudMatrix 384 → 160,000 accelerator cards, with Huawei reporting sold more than 300 systems and serving 20+ customers.
- Technical advantages & claimed multiples: designs emphasize full optical interconnects, liquid cooling and high per‑node bandwidth; Huawei’s Atlas 950 is reported to outperform a comparable NVL144 by about 56.8× card count, 6.7× total compute, 15× memory capacity, and 62× interconnect bandwidth.
- Operational challenges & supply‑chain impact: very high density drives per‑cabinet power consumption commonly >100 kW, requiring upgraded power delivery and liquid cooling; this creates opportunities for NIC makers, optical module and connector suppliers, cooling specialists and system integrators/software.
- Alibaba Cloud: Panjiu 128 superpod with up to 128 AI chips, 50% higher inference performance.
- Huawei: CloudMatrix 384 (up to 160k cards cascaded), Atlas 950 (8,192 cards), Atlas 960 (15,488 cards); sold >300 systems to 20+ customers.
- Inspur: SD200 “YuanNao” for trillion-parameter training.
- Muxi: Yaolong optical-interconnect designs, high-density liquid-cooled solutions.
- Baidu: Baige platform v5.0 with Kunlun-chip superpods.

superpods are the new battleground for AI infrastructure in China.
Overview — why superpods matter for AI infrastructure
Artificial intelligence (AI) is reshaping industries at unprecedented speed.
That transformation depends on massive compute capacity and practical system design.
As model parameters scale from hundreds of millions to trillions, the industry is moving beyond single‑server and traditional cluster approaches.
Superpods (超节点, Superpod) are increasingly becoming the foundation for large‑scale AI workloads.

Resume Captain
Your AI Career Toolkit:
- AI Resume Optimization
- Custom Cover Letters
- LinkedIn Profile Boost
- Interview Question Prep
- Salary Negotiation Agent

What is a “superpod” (超节点 Superpod)?
A superpod is a system architecture that integrates thousands of accelerators (GPU/AI chips) into a single logical compute unit.
The term was popularized by NVIDIA (Yingweida 英伟达, NVIDIA) and describes tightly coupled racks or cabinets connected with very high‑speed interconnects.
Superpods behave more like a single, massive compute node than a loose collection of servers.
Compared with traditional server‑based clusters, superpods use ultra‑high‑bandwidth, low‑latency interconnects to reduce server‑to‑server bandwidth and latency limits.
The result is substantially better system-level compute efficiency for both training and inference of very large models.

Recent product moves in China — Alibaba Cloud and Huawei
At the 2025 Yunqi (Cloud) Conference, Alibaba Cloud (Aliyun 阿里云, Alibaba Group 阿里巴巴) unveiled the Panjiu 128 (磐久128) superpod AI server.
The Panjiu 128 reportedly integrates Alibaba’s in‑house CIPU 2.0 chip and high‑performance EIC/MOC NICs.
The Panjiu 128 supports up to 128 AI compute chips per cabinet.
Alibaba claims inference performance up to 50% higher than comparable traditional architectures at the same raw compute level.
Huawei (Huawei 华为) has pursued a larger‑scale approach with its CloudMatrix 384 superpods.
Earlier in 2025 Huawei described building clusters with more than ten thousand accelerators.
Huawei says it’s possible to cascade 432 CloudMatrix 384 superpods into a mega‑cluster supporting up to 160,000 accelerator cards (a “160k‑card” megacluster).
Huawei claims that such a megacluster can train trillion‑parameter and ten‑trillion‑parameter models.
At Huawei’s Full‑Connection Conference in September 2025, the company said it had sold more than 300 CloudMatrix 384 superpod systems.
Huawei also said it was serving over 20 enterprise and government customers with those systems.
Huawei announced Atlas series superpods including Atlas 950 SuperPoD (8,192‑card scale) targeted for Q4 2026.
Huawei also announced a next‑generation Atlas 960 SuperPoD at 15,488 cards aimed for Q4 2027.

Find Top Talent on China's Leading Networks
- Post Across China's Job Sites from $299 / role, or
- Hire Our Recruiting Pros from $799 / role
- Qualified Candidate Bundles
- Lower Hiring Costs by 80%+
- Expert Team Since 2014
Your First Job Post

Other domestic players and product lines
- Inspur (Langchao 浪潮, Inspur) released the SD200 “YuanNao” (元脑SD200) superpod AI server aimed at trillion‑parameter model training.
- Muxi (沐曦股份) and partners have showcased optical‑interconnect superpods, 3D Mesh “Yaolong” (耀龙) designs, Shanghai Cube high‑density liquid‑cooled cabinets, and high‑density liquid‑cooled compute POD solutions.
- Baidu (Baidu 百度) announced the Baige (百舸) AI computing platform v5.0 and launched Kunlun‑chip‑based superpods for its cloud offering.

ExpatInvest China
Grow Your RMB in China:
- Invest Your RMB Locally
- Buy & Sell Online in CN¥
- No Lock-In Periods
- English Service & Data
- Start with Only ¥1,000

Why the industry is shifting from single‑card performance to system efficiency
Market observers including Hualong Securities (Hualong Zhenquan 华龙证券) and Minsheng Securities (Minsheng Zhengquan 民生证券) argue that China’s AI competition is focused on system‑level efficiency.
The shift is driven by several practical forces:
- Model scale: Large language models and multi‑modal models require coordinated memory, bandwidth, and compute across many accelerators.
- Engineering pragmatism: Cluster‑level optimization, software stacks, and delivery (engineeringized solutions) matter as much as raw silicon speed.
- Supply‑chain leverage: Higher superpod adoption benefits optical interconnect, cooling, and power‑supply suppliers and prompts upstream investment.
This means investors and operators are starting to value integrated solutions that deliver predictable throughput and reliability at scale.

Technical advantages claimed by leading superpods
Design choices often include full optical interconnects between cabinets, high reliability, high per‑node bandwidth, and low latency.
For example, Huawei describes its Atlas 950 as using an orthogonal architecture with zero‑cable electrical interconnect.
Huawei also highlights improved liquid‑cooling integration that it says doubles the reliability of optical modules under liquid‑cooling conditions.
Huawei’s public comparisons position Atlas 950 well ahead of some competing designs (for example NVIDIA’s announced NVL144 product planned for the second half of next year).
Reported comparative multiples between Atlas 950 and NVL144 include the following:
- Card count: Atlas 950 ~56.8× NVL144
- Total compute (aggregate): Atlas 950 ~6.7× NVL144
- Memory capacity: Atlas 950 ~15× NVL144
- Interconnect bandwidth: Atlas 950 ~62× NVL144

Infrastructure and operational challenges for superpods
Higher per‑cabinet compute density creates major thermal and power challenges.
Multiple vendors’ superpod cabinets — including Huawei CloudMatrix 384 and NVIDIA GB200 NVL72 referenced in industry commentary — now show per‑cabinet power consumption commonly exceeding 100 kW.
As power and compute density grow exponentially, data centers must upgrade in several areas:
- Power delivery: higher‑capacity PDUs, transformers, and redundancy planning.
- Thermal control: liquid cooling and new heat‑dissipation architectures to enable stable operation at high densities.
- Interconnect and reliability engineering: optical modules, routing, and error recovery at scale.
When fully liquid‑cooled superpods are deployed, vendors expect substantial bandwidth and compute improvements.
Those gains require redesigning data‑center electrical and cooling subsystems before operators can realize them reliably.

Investment implications and supply‑chain effects
Investment banks such as Guojin Securities (Guojin Zhengquan 国金证券) see domestic superpod platforms as leaders on key metrics like compute, bandwidth, and memory.
Analysts view superpods as potential catalysts for accelerating deployment of local compute infrastructure.
As superpod penetration increases, demand should lift the optical interconnect, liquid‑cooling, and high‑density cabinet supply chains.
For investors and operators, the shift to system‑level solutions means opportunity beyond chip vendors.
There is explicit upside for NIC makers, optical connector and module suppliers, rack integrators, cooling specialists, and system software vendors who can deliver stable, scalable deployments.

Bottom line — what this means for China AI and global infrastructure
The next phase of the AI infrastructure race is about scale and system integration.
For ultra‑large models, focusing on single‑card benchmarks is giving way to emphasis on end‑to‑end system throughput, reliability, and energy efficiency.
In China, hyperscalers and local hardware firms are betting that well‑engineered superpods — combined with open ecosystems and industrialized delivery — offer the fastest route to competitive, sovereign AI compute capacity.
superpods are the lever that can unlock large‑model performance and the broader supply‑chain growth that follows.
