Amazon used its annual AWS re:Invent conference in Las Vegas to confirm two new families of AI focused chips.

Amazon Web Services announced the next generation of two AWS-designed chip families – AWS Graviton4 and AWS Trainium2 – which AWS said would deliver “advancements in price performance and energy efficiency for a broad range of customer workloads, including machine learning (ML) training and generative artificial intelligence (AI) applications.”

It comes after AWS and Nvidia announced “an expansion of their strategic collaboration to deliver the most advanced infrastructure, software, and services to power customers’ generative artificial intelligence (AI) innovations.”

AWS Graviton4 and AWS Trainium2 (prototype).
Image credit Business Wire

New chips

AWS said that the new Graviton4, which is based on an ARM architecture, provides up to 30 percent better compute performance, 50 percent more cores, and 75 percent more memory bandwidth than current generation Graviton3 processors, which it said will deliver improved price performance and energy efficiency for workloads running on Amazon EC2.

Meanwhile Amazon’s own Trainium2 is designed to deliver up to 4x faster training than first generation Trainium chips and will be able to be deployed in EC2 UltraClusters of up to 100,000 chips. AWS said this will make it possible to train foundation models (FMs) and large language models (LLMs) in a fraction of the time, while improving energy efficiency up to 2x.

“Silicon underpins every customer workload, making it a critical area of innovation for AWS,” said David Brown, vice president of Compute and Networking at AWS. “By focusing our chip designs on real workloads that matter to customers, we’re able to deliver the most advanced cloud infrastructure to them.”

“Graviton4 marks the fourth generation we’ve delivered in just five years, and is the most powerful and energy efficient chip we have ever built for a broad range of workloads,” said Brown. “And with the surge of interest in generative AI, Tranium2 will help customers train their ML models faster, at a lower cost, and with better energy efficiency.”

Thousands of customers

At the moment, AWS offers more than 150 different Graviton-powered Amazon EC2 instance types and has built more than 2 million Graviton processors, and has more than 50,000 customers.

EC2 UltraClusters of Trainum2 are designed to deliver the highest performance, most energy efficient AI model training infrastructure in the cloud.

All of this is aided after AWS said it will offer access to Nvidia’s latest H200 AI graphics processing units.

And as part of its deepening relationship with Nvidia, AWS said it will operate more than 16,000 Nvidia GH200 Grace Hopper Superchips, which contain Nvidia GPUs and Nvidia’s ARM-based general-purpose processors.

Tom Jowitt

Tom Jowitt is a leading British tech freelancer and long standing contributor to Silicon UK. He is also a bit of a Lord of the Rings nut...

Recent Posts

Tesla To Ask Shareholders To Reinstate Elon Musk’s $56 Billion Package

Tesla shareholders to be asked to reinstate Elon Musk's $56 billion pay package, days after…

7 hours ago

Telegram To Reach One Billion Users Within Year

Catching WhatsApp? Billionaire founder of Telegram claims encrypted platform will reach one billion users within…

8 hours ago

Judge Dismisses Some Harm Claims Against Meta, Zuckerberg

Good news for Mark Zuckerberg as judge dismisses some claims in dozens of lawsuits alleging…

9 hours ago

Google Begins Removal Of California News Ahead Of Proposed Law

Consequences of Assembly Bill 886. Google begins removing California news websites from some search results

10 hours ago

Tim Cook Says Apple Considering Factory In Indonesia

CEO Tim Cook during visit to Jakarta says Apple will look into building a manufacturing…

11 hours ago

Canada To Implement Digital Services Tax This Year

Introduction of digital services tax on tech firms will begin in 2024 Canadian government confirms,…

15 hours ago