Sitemap
Predict

where the future is written

Member-only story

Trainium and Inferentia: Amazon Takes On NVIDIA

4 min readMar 24, 2025

--

Photo by on

Amazon’s Trainium chip has been making headlines lately. Amazon Web Services (AWS) has been aggressively pitching this chip to customers, offering server rentals with computing power comparable to NVIDIA’s H100 GPUs but at a steep discount (posts on X suggest it’s just a quarter of what NVIDIA charges).

It’s a bold move, and there’s a mix of strategy, economics, and market dynamics driving it.

NVIDIA’s H100 GPUs are, as you might know, the go-to for AI training, powering everything from large language models to image/video generation. But they come with a hefty price tag. Last I checked, running an EC2 P5.48xlarge instance costs around $98.32/hour (on-demand pricing). Sure, it’s a beast of a setup, with 8 H100 GPUs, 640 GB of VRAM, and 2 TB of RAM, but that price still stings.

Besides, they’re always short on supply.

The supply shortage is well-documented. TSMC’s 4nm production process just can’t keep up with the crazy demand for H100s. And it’s not for lack of effort. Yields (the percentage of chips that actually work) take a serious hit…

Predict
Predict
Ashraff Hathibelagal
Ashraff Hathibelagal

Written by Ashraff Hathibelagal

Computer Scientist | Writer | Effective Accelerationist

Responses (5)