Gadgets

Amazon Launches Trainium 3, Promises 4× Faster AI Training and Lower Energy Use

In a packed keynote at AWS reInvent 2025, Amazon Web Services unveiled Trainium 3, its latest custom silicon built to accelerate AI training and inference at enterprise scale. The announcement, which lifted AWS stock by 2.3 percent in after hours trading, promises more than four times the performance of its predecessor while consuming 40 percent less power.

It is a combination that could significantly reduce the cost of building and operating the large data centers driving the global AI surge. With worldwide AI spending projected to reach 200 billion dollars in 2025 according to Gartner, Trainium 3 positions AWS not only as a cloud provider but as a serious hardware contender challenging NVIDIAs dominance.

Trainium 3 will run inside AWSs new UltraServer architecture, where each server contains 144 of the new chips. These servers can scale to clusters comprising as many as one million chips, creating the capacity to train models on the scale of GPT 4 in weeks instead of months. Early performance benchmarks shared during the keynote show Trainium 3 delivering 4.2 times faster training on large language models compared to Trainium 2 while consuming substantially less electricity.

It makes AI accessible to more businesses without overwhelming financial or energy costs, AWS Chief Executive Matt Garman said. The remarks come at a time when AI data centers could consume up to eight percent of global electricity by 2030 according to estimates from the International Energy Agency.

The chips appear to excel in high scale workloads. Anthropic, one of the earliest adopters, reported a 3.5 times speed improvement on its Claude models. Startups such as Adept.ai noted a thirty five percent reduction in inference costs for real time AI agents. AWS also highlighted Trainiums sustainability advantages, a timely focus as regulators including those behind the European Unions Green Deal intensify scrutiny of AIs carbon footprint.

One training run of GPT 3 is estimated to consume 1287 megawatt hours of energy according to researchers at the University of Massachusetts. We are not just faster. We are greener, Garman said, linking the chip to AWSs carbon neutral pledge.

The launch reflects a broader strategic shift toward hybrid AI infrastructure. AWS confirmed that Trainium 4, scheduled for 2026, will support NVIDIAs NVLink Fusion interconnect to enable direct integration with GPU systems. The move signals cooperation rather than confrontation with the GPU manufacturer which currently controls about 80 percent of the AI accelerator market. The approach mirrors the success of AWSs Graviton processors, which captured roughly 25 percent of cloud workloads by 2025 due to their cost savings.

However, AWS still faces challenges. NVIDIAs H100 and H200 GPUs remain the leaders in raw performance, and Trainium 3 trails in certain multimodal tasks, performing around 15 percent slower on vision models according to MLPerf benchmarks. Supply limitations may also slow adoption, as AWS intends to prioritize internal use of the chips. Enterprise availability is not expected before the second quarter of 2026. Critics, including Forrester analyst Mike Gualtieri, argue that Trainium may be excellent for AWS customers but will be difficult to obtain without entering deeper cloud dependency.

With NVIDIA facing antitrust scrutiny and growing pressure on power grids worldwide, Amazon’s energy efficient challenger could shift the industry landscape, provided it delivers the scalability and reliability enterprises require.