Newsvidia

AI Chips Discounts – Amazon undercuts Nvidia pricing by 25%, leveling market for simpler inferencing tasks

ai chips amazon nvidia pricing

AI Chips Discounts – Amazon undercuts Nvidia pricing by 25%, leveling market for simpler inferencing tasks

Servers powered by AWS’ Trainium chip could democratize access for enterprises looking to experiment with AI, but who can’t get access to or afford Nvidia’s infrastructure.

Amazon is looking to take on Nvidia, the dominant player in the AI market, by purportedly undercutting its pricing by 25%.

An anonymous AWS cloud customer has revealed that the cloud giant pitched it on renting servers powered by its purpose-built Trainium chips, offering the same computing power as Nvidia’s highly sought-after H100 chips at three-quarters of the price.

Coincidentally or not, the news came out as Nvidia hosts its flagship event, Nvidia GTC, this week.

Forrester Senior Analyst Alvin Nguyen, said:

By offering access to their Trainium chips, Amazon is mitigating the constraint of access to GPUs for AI and keeping interest in AI in the cloud,

“This also mitigates their dependency on Nvidia as a supplier for a critical service.”

The most focused and aggressive of the large CSPs

Nvidia’s architecture is highly sought after, but expensive and difficult to come by. Thus, the company is facing growing competition, even from its own customers, including Google, which is working on the next generation of its Tensor Processing Units to rival GPUs, Microsoft, which offers its custom-designed Azure chips, and, of course, Amazon.

Other players such as AMD, Broadcom, Qualcomm, Intel, and Marvell have also taken aim at Nvidia.

All of the large cloud service providers (CSPs) have developed dedicated silicon, designed in house, to offer as an alternative to commercially available chips at a “healthy discounted rate,” explained Matt Kimball, VP and principal analyst for data compute and storage at Moor Insights & Strategy.

For Amazon, that’s the Trainium, the Graviton processor, and the Inferentia chip. AWS has been able to monetize its homegrown chips quite effectively and has been “wildly successful” with Graviton, he noted. It has been the most focused and aggressive of the large CSPs in chasing the enterprise AI market with its own silicon.

Kimball, said:

What AWS is doing is smart — it is telling the world that there is a cost-effective alternative that is also performant for AI training needs,

“It is inserting itself into the AI conversation.”

AI inferencing geared to the everyday enterprise

While Nvidia continues to raise the bar on AI performance with its latest chips, it’s like a set of golden handcuffs, said Nguyen. That is, the company doesn’t have much incentive to create lower-end chips for smaller workloads. This creates an opportunity for Amazon and others to gain footholds in these adjacent markets.

Nguyen, said:

This serves to fill underneath,

“Purpose-built training chips are a nice to have for the open market.”

These chips can make chain-of-thought (CoT) and reasoning models, in which AI systems think out each step before providing an output, more accessible to more companies, essentially democratizing AI, he said.

Kimball agreed,

AWS will find a strong interest in the enterprise that is new to high performance computing and AI: the commercial customer,

The result may not necessarily be production-ready AI, Nguyen emphasized, but it at least allows enterprises the ability to experiment. Many haven’t been able to do so due to access issues; one company he consults with, for instance, had to wait 18 months to even get access to an Nvidia GPU.

Nguyen noted,

That’s a long time for any business to wait,

Conversation-starter or vendor lock-in?

Training cost is highly dependent on many factors — the number of GPUs, token count, model size, and frameworks — all of which need to come into play when enterprises consider one chip over another, Kimball emphasized.

This move from Amazon will undoubtedly be a conversation starter, and the AWS AI stack can be a great fit for some enterprises “due to its (relative) ease of use, cost and overall performance,” he said.

However, enterprises used to working with Nvidia’s compute unified device architecture (CUDA) need to think about the cost of switching to a whole new platform like Trainium. Furthermore, Trainium is only available on AWS, so users can get locked in.

Kimball noted,

Whereas,

That said, however,

I can shop for the best price on the Nvidia front.

He said,

I’m still locked in to Nvidia, and from a capex perspective, it is more costly,

Nvidia will still dominate for the foreseeable future

Analysts agree that, despite increased competition, Nvidia will remain on top.

Mario Morales, group VP for enabling technologies and semiconductors at IDC, said:

AWS can provide better performance efficiency, just like others who are targeting AI inferencing at a lower cost,

However, he said, “Nvidia is launching next-generation architectures today that no other company, including AWS, comes close to in terms of overall performance.”

Nguyen agreed,

No one’s displacing them anytime soon. At this point, Nvidia is doing everything right; they’re controlling everything they can control,

He said,

They have the top fabs. You can’t get a GPU that’s more powerful.

READ the latest news shaping the Nvidia market at Newsvidia

AI Chips Discounts – Amazon undercuts Nvidia pricing by 25%, leveling market for simpler inferencing tasks, source

Follow us on LinkedIn!

Join our weekly newsletter!

Please enable JavaScript in your browser to complete this form.

Your Header Sidebar area is currently empty. Hurry up and add some widgets.