Insights Business| SaaS| Technology Why Chinese Tech Companies Ordered Two Million Nvidia H200s Despite Tariffs and Regulatory Risk
Business
|
SaaS
|
Technology
Apr 16, 2026

Why Chinese Tech Companies Ordered Two Million Nvidia H200s Despite Tariffs and Regulatory Risk

AUTHOR

James A. Wondrasek James A. Wondrasek
Graphic representation of Chinese tech companies ordering Nvidia H200 GPUs despite tariffs and regulatory risk

Chinese tech companies — Alibaba, ByteDance, Tencent, and Baidu — have placed orders for more than two million Nvidia H200 GPUs. At $27,000 per chip, that’s a $54 billion-plus commitment before the 25% tariff. Full upfront payment. No cancellations. And delivery isn’t guaranteed — Beijing’s customs authority has already told logistics firms to pause port clearance for H200 chips.

So why are they doing it?

Three structural forces make this demand commercially unavoidable: the performance ceiling imposed by export controls, the impossibility of domestic substitution at scale, and the competitive cost of falling further behind in frontier AI training. It also matters beyond China — the 2M+ unit order backlog is competing directly against available H200 inventory and shaping delivery timelines for every organisation buying AI compute in 2026. For the full procurement planning framework, see our AI GPU procurement hedge strategies guide for CTOs.


What Did China’s Tech Sector Actually Order — and How Much Did It Cost?

Two million H200 units. At $27,000 per chip, that is a $54 billion-plus market commitment — before the 25% tariff. Alibaba and ByteDance reportedly placed individual orders of 200,000+ units each. ByteDance’s broader AI hardware spending for 2026 is estimated at approximately $14 billion. That is not a procurement line item. That is a capital expenditure with strategic significance.

The tariff arithmetic is actually quite revealing. With the 25% tariff applied, an eight-chip H200 module costs approximately 1.5 million yuan through official channels. Before BIS policy was clarified, grey-market H200 modules were trading at over 1.75 million yuan. The tariff-inclusive official price is cheaper than what Chinese buyers were paying informally. The tariff creates a cost increase; it does not create the demand.

Against 2 million units ordered, Nvidia holds approximately 700,000 units of available H200 inventory. Fulfilling the remaining backlog requires new TSMC CoWoS-S production runs — more than three months from wafer start to delivered unit. The China General Administration of Customs has also issued a directive instructing logistics companies not to submit clearing requests for H200 chips at Chinese ports, leaving physical delivery on hold.


Why Is the H200 the Ceiling — and What Happened to H20 Exports?

The Nvidia H200 is the highest-performance chip Nvidia can legally export to China under current Bureau of Industry and Security (BIS) rules.

Above it: Nvidia Blackwell (B100, B200, GB200 NVL72). Blackwell requires TSMC’s 4NP process node, which is not available for China-bound chip production under US national security requirements. No tariff, no licensing pathway, no political arrangement makes Blackwell available to Chinese buyers.

Below it: the Nvidia H20 — Nvidia’s China-specific downgraded variant, engineered with reduced bandwidth to stay within earlier export control thresholds. The Trump administration imposed a presumption of denial for H20 exports in April 2025, costing Nvidia an estimated $10 billion or more in foregone sales. That demand vacuum channelled into H200 orders once BIS case-by-case review created a legal pathway.

The performance gap between H20 and H200 is not incremental. The H200 offers six times the floating-point performance of the H20, 50% more HBM3e memory capacity, and 20% higher memory bandwidth. The H20 was deliberately designed to be inferior — that was the point. But that inferiority is precisely why it fails for frontier AI training. The H200 is China’s only option at Hopper-class performance.

For detail on the BIS approval framework, see how the 25% tariff and BIS case-by-case review actually works.


Why Can the Huawei Ascend 910C and Domestic Alternatives Not Fill the Gap?

The Huawei Ascend 910C is China’s most capable domestically produced AI accelerator as of 2026 — manufactured by SMIC on a 7nm process node. It is real, production-grade hardware. But for large model training, the performance gap relative to the H200 is decisive.

DeepSeek researchers working directly with Ascend hardware concluded the 910C performs approximately 60% as well as the Nvidia H100. The key constraint is not Huawei’s design capability; it is SMIC’s process node ceiling. US export restrictions prevent SMIC from accessing ASML EUV machines, locking it at 7nm. TSMC’s 4N node, used for the H200, is approximately two process generations more advanced. The last Nvidia chip manufactured at 7nm was the A100, released in 2020.

That gap is not narrowing. TrendForce projects Nvidia’s best AI chips will be seventeen times more powerful than Huawei’s best by the second half of 2027. The successor Ascend 960 is not expected until Q4 2027. Chinese AI labs cannot wait two-plus years for a chip that will still be catching up on arrival.

Below Huawei Ascend sit Cambricon, Moore Threads, and Biren — even further behind for frontier training workloads.


What Is CUDA Lock-in — and Why Does It Make Domestic Substitution Commercially Impossible?

CUDA — Compute Unified Device Architecture — is Nvidia’s proprietary parallel computing platform. And it is not just a programming language. It is a comprehensive ecosystem of libraries (cuDNN, cuBLAS, NCCL) and community tools built over nearly two decades. Roughly 75% of AI training chips in Chinese data centres run on CUDA-dependent workloads, despite years of US export restrictions. That is the accumulated reality of how the global AI software stack was built.

Switching from CUDA to Huawei’s CANN (Compute Architecture for Neural Networks) is not a configuration change. It requires rewriting training code at the library level, revalidating model performance, and retraining engineering teams. Industry analysts estimate 6–18 months of effort to port a mature CUDA workload to CANN — during which every competing lab running H200s continues to train.

For a leading Chinese AI lab, moving to Ascend hardware means accepting a ~40% hardware performance deficit and a software transition penalty simultaneously. Two compounding losses that make H200 procurement economically rational even at tariff-inclusive pricing. Western CTOs evaluating their own GPU procurement strategy amid this demand competition will find the AI infrastructure planning framework for semiconductor geopolitics a useful parallel lens.


Why Is Nvidia Demanding Full Upfront Payment — and What Does It Mean for Chinese Buyers?

Nvidia’s payment terms for Chinese H200 buyers: 100% payment at order placement, no cancellation rights, no configuration changes after order.

If Beijing blocks import clearance after payment, Nvidia retains the revenue. The regulatory risk is transferred entirely to the Chinese buyer. Nvidia has effectively monetised the regulatory uncertainty rather than absorbed it.

From the Chinese buyer’s perspective, though, the calculation is not unreasonable. The competitive stakes of a six-month delay in receiving H200s are higher than the capital risk of a prepayment that may not clear customs. Both regulatory gates — Washington’s BIS approval and Beijing’s import clearance — are open simultaneously, and neither is guaranteed. You pay now or you fall behind. That’s the choice on the table.


What Is Beijing Doing About It — Approvals, Rejections, and the Domestic Bundling Rule?

Beijing’s regulatory position is more nuanced than either a straightforward approval or a blanket ban — and that ambiguity appears to be deliberate.

H200 imports will be permitted for selected commercial-sector buyers. Alibaba Cloud, ByteDance, and Baidu are in scope. Military entities, critical infrastructure operators, and state-owned enterprises are not. Meanwhile, customs authorities told agents H200s were not permitted to enter ports, while separately instructing domestic tech companies not to purchase H200 chips “unless necessary” — ambiguity without legal clarity, which lets Beijing calibrate restrictions without a formal instrument that could be cited or challenged.

The domestic chip bundling requirement is Beijing’s attempt to have it both ways. Companies importing H200s are required to also purchase a defined ratio of domestically produced chips — from Huawei Ascend, Cambricon, and Moore Threads. Beijing knows domestic alternatives cannot substitute for frontier training, but it is preserving demand for domestic producers rather than letting H200 access displace them entirely.

On the US side, the BIS final rule published January 15, 2026 shifts H200 export licence review to case-by-case — subject to a 25% tariff, a 50% volume cap, KYC verification, third-party performance testing in the US before export, and physical security requirements at the end-user site.

The result: an H200 order requires BIS approval from Washington and import clearance from Beijing, with full payment already made to Nvidia before either gate fully opens.

For BIS approval detail, see how the 25% tariff and BIS approval process works. For legislation targeting offshore GPU training workarounds, see the AI Overwatch Act and Remote Access Security Act.


What Does China’s H200 Demand Mean for GPU Availability for Everyone Else in 2026?

Two million units ordered. Approximately 700,000 units of available H200 inventory. Even if only a fraction of Chinese orders are fulfilled, that volume consumes a meaningful share of constrained global supply.

The binding constraint is not Nvidia’s willingness to sell. It is TSMC’s CoWoS advanced packaging capacity, which is approximately three times short of customer demand per TSMC CEO C.C. Wei. Advanced packaging is not a bottleneck that money resolves — it requires physical fab capacity that takes 12–24 months to expand.

HBM3e memory is equally constrained. SK Hynix controls roughly 60% of the global HBM market and is sold out through 2026. Micron‘s HBM production is fully committed through 2026. Three suppliers, all sold out.

If you are buying AI compute outside China in 2026, the geopolitical story of Chinese H200 demand is also your procurement planning story. The 2M-unit backlog does not need to be fully fulfilled to affect your delivery timelines and pricing — it only needs to exist and compete for the same constrained inventory. For HBM and packaging bottleneck analysis, see HBM4 delays and GDDR7 shortages — the memory bottlenecks squeezing GPU supply. For a full framework on AI infrastructure procurement planning amid China demand competition, see our full CTO planning framework for AI chip tariffs and semiconductor geopolitics.


Frequently Asked Questions

Why did Chinese tech companies order 2 million Nvidia H200 GPUs when exports were still uncertain?

Waiting was competitively worse. CUDA lock-in makes domestic substitution commercially infeasible without an 18-month-plus engineering transition. The H200 is the only hardware at Hopper-class performance accessible to Chinese buyers. The competitive cost of delaying AI training outweighs the financial risk of full prepayment.

What is the performance difference between the Nvidia H200 and the Huawei Ascend 910C?

The Ascend 910C delivers approximately 60% of H100 performance in real-world conditions, per DeepSeek researchers. The H200 provides 50% more HBM3e memory capacity and 20% higher bandwidth. The gap is projected to widen to 17x by the second half of 2027.

What is CUDA lock-in and why does it matter for AI chip decisions?

CUDA is Nvidia’s proprietary computing platform — a comprehensive ecosystem of libraries (cuDNN, cuBLAS, NCCL) and community tools built over nearly two decades. Migrating a production training workload to Huawei’s CANN or AMD’s ROCm takes an estimated 6–18 months of engineering effort. Competitors on H200s keep training the whole time.

Can Chinese companies still buy Nvidia GPUs after the 2025 tariffs and export controls?

Yes, under specific conditions. The BIS case-by-case review replaced the blanket presumption of denial for H200 exports to China. Conditions include a 25% tariff, a 50% volume cap, KYC verification, third-party performance testing in the US, and end-user physical security requirements. Commercial-sector buyers are in scope; military, SOE, and critical infrastructure entities are not. Beijing import approval is required independently of BIS.

What is the difference between the US tariff on H200 chips and the export controls?

Two distinct mechanisms. The 25% tariff raises the price but does not prohibit purchase. BIS licensing determines whether a sale can happen at all. A Chinese company can pay the tariff and still be denied an export licence. Both apply simultaneously.

Why can’t China simply produce its own AI chips to replace Nvidia H200s?

SMIC is constrained to 7nm due to US restrictions on advanced lithography equipment — specifically, ASML EUV machines. The H200 is manufactured on TSMC’s 4N node, approximately two process generations more advanced. No design refinement overcomes a process node ceiling.

Why is Nvidia requiring full upfront payment from Chinese H200 buyers?

Full upfront payment transfers regulatory risk from Nvidia to the buyer. If Beijing blocks import clearance after payment, Nvidia retains the revenue. For Chinese buyers, full prepayment is preferable to having no H200 access at all.

What is the Domestic Chip Bundling Requirement that Beijing proposed?

Beijing proposed requiring companies importing H200s to also purchase a defined ratio of domestically produced AI chips — from Huawei Ascend, Cambricon, and Moore Threads — alongside each H200 order. The policy sustains domestic chip industry revenue while allowing limited H200 access for commercial AI labs. It had not been enacted at time of reporting.

What does the 50% BIS volume cap mean in practice for H200 supply to China?

Aggregate H200 and AMD MI325X shipments to China cannot exceed 50% of the volume shipped to US customers in the same period. If Nvidia ships 100,000 H200 units to US customers in a quarter, a maximum of 50,000 can go to China. Given that US demand is itself constrained by TSMC capacity, this cap creates a hard ceiling on Chinese order fulfilment independent of payment or regulatory status.

Are other Western GPU buyers affected by China’s H200 demand?

Yes, directly. The 2M+ Chinese order backlog competes against approximately 700,000 units of available H200 inventory. TSMC’s CoWoS packaging capacity is severely constrained through mid-2026, and SK Hynix is sold out of HBM through 2026. Any H200 units fulfilling Chinese orders are not available for Western buyers.

What happened to Nvidia’s H20 chip sales to China before the H200 deal?

Nvidia sold more than 1 million H20 chips in China in 2024. The Trump administration imposed a presumption of denial for H20 exports in April 2025, costing Nvidia an estimated $10 billion-plus in foregone sales. That demand vacuum channelled into H200 orders once BIS created a legal pathway.

What are Nvidia Blackwell chips and why can’t China get them?

Nvidia Blackwell (B100, B200, GB200 NVL72) is Nvidia’s next-generation GPU architecture, substantially more capable than the H200. Chinese buyers cannot access any Blackwell-class hardware — it requires TSMC’s 4NP process node, which is not available for China-bound chip production under US national security requirements. No tariff or licensing arrangement changes this.

AUTHOR

James A. Wondrasek James A. Wondrasek

SHARE ARTICLE

Share
Copy Link

Related Articles

Need a reliable team to help achieve your software goals?

Drop us a line! We'd love to discuss your project.

Offices Dots
Offices

BUSINESS HOURS

Monday - Friday
9 AM - 9 PM (Sydney Time)
9 AM - 5 PM (Yogyakarta Time)

Monday - Friday
9 AM - 9 PM (Sydney Time)
9 AM - 5 PM (Yogyakarta Time)

Sydney

SYDNEY

55 Pyrmont Bridge Road
Pyrmont, NSW, 2009
Australia

55 Pyrmont Bridge Road, Pyrmont, NSW, 2009, Australia

+61 2-8123-0997

Yogyakarta

YOGYAKARTA

Unit A & B
Jl. Prof. Herman Yohanes No.1125, Terban, Gondokusuman, Yogyakarta,
Daerah Istimewa Yogyakarta 55223
Indonesia

Unit A & B Jl. Prof. Herman Yohanes No.1125, Yogyakarta, Daerah Istimewa Yogyakarta 55223, Indonesia

+62 274-4539660
Bandung

BANDUNG

JL. Banda No. 30
Bandung 40115
Indonesia

JL. Banda No. 30, Bandung 40115, Indonesia

+62 858-6514-9577

Subscribe to our newsletter