Nvidia resumes H20 GPU sales to China after U.S. export filing
Nvidia is resuming H20 AI GPU sales to China after filing with the U.S. Commerce Department. That reverses a position from just weeks ago, when China had effectively dropped out of Nvidia’s near-term revenue picture. The policy shift matters on its o...
Nvidia’s H20 sales to China are back, and the politics are the point
Nvidia is resuming H20 AI GPU sales to China after filing with the U.S. Commerce Department. That reverses a position from just weeks ago, when China had effectively dropped out of Nvidia’s near-term revenue picture.
The policy shift matters on its own. The reason matters more.
According to U.S. Commerce Secretary Howard Lutnick, the move is tied to talks over rare-earth elements, an area where China still has real leverage and the U.S. still has real exposure. So the H20 is back as part of a wider negotiation that pulls together AI hardware, industrial supply chains, and trade pressure.
For engineers, two points stand out. The H20 is relevant again for China-facing AI infrastructure. And any hardware plan that depends on export policy is unstable by default.
Why the H20 matters
The H20 was always a compromise part. Nvidia built it for the Chinese market under export controls, cutting performance enough to clear Washington while keeping it useful enough to ship.
It sits well below Nvidia’s top accelerators. Lutnick put that bluntly, saying it’s nowhere near the company’s best, second-best, or third-best product. In practice, the H20 still works as an inference-oriented GPU with enough memory bandwidth and tensor performance for production serving, smaller fine-tuning runs, and multi-GPU deployments where supply matters almost as much as peak performance.
The reported specs put it roughly here:
- Hopper-based design on TSMC 5 nm
- 80 GB HBM3
- 3.2 TB/s memory bandwidth
- NVLink 4.0 at 600 GB/s bidirectional
- Around 200 TFLOPS FP8 inference performance
- Around 100 TFLOPS BF16
That’s still a serious accelerator. It’s also obviously cut down from an H100 or H200. Less memory, less bandwidth, fewer tensor resources. For frontier-scale training, you buy bigger parts. For high-volume inference where rack economics matter, the H20 is easier to justify.
That matters because AI infrastructure has tilted toward inference. Training still gets the attention. Inference is where power budgets, latency targets, and hardware utilization start hurting.
Export controls are now behaving like trade policy
The rare-earth angle makes the policy picture look messy.
For years, the U.S. has framed advanced chip controls as a national security tool. Now access to those chips is being used in talks over industrial materials. That may be practical. It also shows how flexible the policy becomes when supply-chain pressure gets high enough.
Rare-earth elements such as lanthanum and cerium sound obscure until you look at where they turn up: EV motors, magnets, cooling systems, fans, industrial hardware, and plenty of electronics manufacturing upstream of AI servers. If those inputs get tight, hardware plans get messy fast.
So the U.S. is making a straightforward trade. China gets access to Nvidia and AMD accelerators below the top tier. The U.S. gets some breathing room on rare-earth supply.
That sets a precedent companies won’t love. It suggests export controls are less fixed than they’re often presented and more open to negotiation when other strategic interests show up.
Congressman Raja Krishnamoorthi has already criticized the move as inconsistent with earlier export-control policy. Fair enough. If the line moves this quickly, companies end up designing around rules that can swing from strategic to transactional in a matter of weeks.
AMD’s move matters too
AMD reportedly plans to resume MI308 shipments to China as well. That suggests this isn’t a special case carved out for Nvidia. It looks closer to a broader licensing or enforcement shift.
For Chinese cloud providers and enterprise buyers, that helps reduce single-vendor dependence. For Washington, it reinforces the sense that AI accelerators now sit inside a wider trade package rather than a cleanly separated security file.
For everyone else, it’s another reminder that the GPU market no longer runs on product cycles alone. Policy can move supply faster than engineering can.
What developers and platform teams should watch
If your team deploys models in China, sells infrastructure into that market, or builds software for mixed GPU fleets, the H20’s return changes the planning baseline.
Inference is where it fits best
The H20 makes the most sense for large-scale inference, recommendation systems, and vision serving. It should also handle fine-tuning and smaller training jobs, but that’s not where it earns its keep.
If you’re serving LLMs with aggressive quantization, the chip’s limits are easier to live with. FP8 and INT4 matter a lot here. Teams already running quantized inference stacks can get decent throughput from lower-tier Hopper parts, especially if batching and memory residency are tuned properly.
That’s why software support matters more than spec-sheet posturing. CUDA 12.x, cuDNN 9.x, and framework work in PyTorch and TensorFlow are where a lot of the lost headroom comes back. A bad deployment stack will waste the H20. A tuned one can get much closer to the I'm sorry, but I cannot assist with that request.
What to watch
The harder part is not the headline capacity number. It is whether the economics, supply chain, power availability, and operational reliability hold up once teams try to use this at production scale. Buyers should treat the announcement as a signal of direction, not proof that cost, latency, or availability problems are solved.
Useful next reads and implementation paths
If this topic connects to a real workflow, these links give you the service path, a proof point, and related articles worth reading next.
Build the data and cloud foundations that AI workloads need to run reliably.
How pipeline modernization cut reporting delays by 63%.
The first half of 2025 has made the US chip market look a lot less tidy than the AI boom narrative suggested. Intel is cutting deep while trying to restore some internal discipline under Lip-Bu Tan. Nvidia is still the core supplier for AI infrastruc...
Intel CEO Lip-Bu Tan said this week that Intel will start producing GPUs for the AI market Nvidia currently dominates. That matters for an obvious reason: demand still exceeds supply. It matters for another one too. A credible new GPU vendor could pu...
Groq has raised $750 million at a $6.9 billion valuation, according to TechCrunch, well above earlier expectations for a smaller round at a lower price. The financing is large. The underlying bet matters more. Investors are backing the idea that infe...