Skip to content
the king of matrix multiplication

For Nvidia, it’s AI or bust as it reports a record-breaking quarter

Everybody wants GPUs for AI, and that’s making Nvidia very happy (and rich).

Benj Edwards | 75
A lone red complete toy robot holding the NVIDIA logo lying on a pile of discarded blue toy robot parts
Credit: Getty Images | Aurich Lawson
Credit: Getty Images | Aurich Lawson
Story text

On top of Wednesday’s news that Nvidia earnings have performed far better than expected, Reuters reports that Nvidia CEO Jensen Huang expects the AI boom to last well into next year. As a testament to this outlook, Nvidia will buy back $25 billion of shares—which happen to be worth triple what they were just before the generative AI craze kicked off.

“A new computing era has begun,” said Huang breathlessly in an Nvidia press release announcing the company’s financial results, which include a quarterly revenue of $13.51 billion, up 101 percent from a year ago and 88 percent from the previous quarter. “Companies worldwide are transitioning from general-purpose to accelerated computing and generative AI.”

For those just tuning in, Nvidia enjoys what Reuters calls a “near monopoly” on hardware that accelerates the training and deployment of neural networks that power today’s generative AI models—and a 60–70 percent AI server market share. In particular, its data center GPU lines are exceptionally good at performing billions of the matrix multiplications necessary to run neural networks due to their parallel architecture. In this way, hardware architectures that originated as video game graphics accelerators now power the generative AI boom.

Nvidia’s most popular AI hardware products currently include A100 and H100 data center GPUs, and Nvidia has also combined the H100 and a CPU into a package called the GH200 “Grace Hopper” chipset that powers Nvidia’s line of computer systems. These are not consumer-grade gaming GPUs like the GeForce RTX 4090—The Verge reports that the H100 chip sells for roughly $40,000—and they can execute a great deal more calculations every second.

Demand for GPUs in AI applications is huge, and Nvidia’s second-quarter data center revenue ($10.32 billion) dwarfed its consumer gaming revenue ($2.49 billion). According to reports from March, OpenAI’s popular AI assistant ChatGPT was projected to utilize as many as 30,000 Nvidia GPUs to run, although exact numbers have not been released by the company. Microsoft is also utilizing data centers full of “tens of thousands” of GPUs to power its implementations of OpenAI’s technology, which it is currently baking into Microsoft Office and Windows 11.

“This is not a one-quarter thing”

NVIDIA's GH200 "Grace Hopper" AI superchip
Nvidia’s GH200 “Grace Hopper” AI superchip.
Nvidia’s GH200 “Grace Hopper” AI superchip. Credit: Nvidia

Nvidia’s market dominance has left competitors like AMD rushing to catch up. But at the moment, Nvidia’s lead seems almost untouchable. In May, Nvidia became the first $1 trillion chip company.

Huang’s move to buy back stock when it is more expensive than ever is risky, but it shows his confidence in Nvidia’s continued success. Demand for its chips has given Nvidia the money needed to pull it off: The firm reported its adjusted gross margins (a financial metric that measures a company’s profitability after accounting for the cost of goods sold) almost doubled to 71.2 percent in its second quarter. Reuters notes that most semiconductor companies have gross margins between 50 and 60 percent.

Huang told Reuters in an interview that two key things are driving Nvidia’s current success: a growing transition from data centers built around CPUs to those built around Nvidia graphics processing units (GPUs), and the rising use of generative AI systems such as ChatGPT.

“These two fundamental trends are what’s behind everything that we’re seeing, and we’re about a quarter into it,” he told Reuters. “It’s hard to say how many quarters are ahead of us, but this fundamental shift is not going to end. This is not a one-quarter thing.”

The case of the billion-dollar FOMO

Of course, with every boom there is the chance of a bust. Every bubble eventually pops—and will Nvidia get caught in the middle of it? Reuters says that some analysts don’t see unlimited demand for Nvidia’s GPU chips. The news agency quotes Dylan Patel of SemiAnalysis as saying that many tech companies are riding the wave of AI hype, speculatively purchasing Nvidia GPUs before figuring out how they can actually make money off of generative AI. Like a billion-dollar case of FOMO.

“They must overinvest in GPUs or risk missing the boat,” Patel told Reuters. “At some point the true use cases will shake out, and many of these players will stop investing, though others will likely continue accelerating investment.”

Then there’s another potential financial snag ahead: product shortages. Reuters says that Huang thinks the biggest risk Nvidia faces is securing supplies necessary to produce its expensive server hardware. The company’s biggest sales success this quarter was its HGX system, a supercomputer built around its H100 GPUs that includes many parts that need to be sourced individually.

“We’re getting great cooperation from our supply chain,” Huang told Reuters in an interview. “And it’s a complicated supply chain. People think it’s a GPU chip. But it’s a very complicated GPU system. It’s 70 pounds. It’s 35,000 components. It’s $200,000.”

In addition, the H100 chips themselves have become difficult to source. Currently, demand for high-powered GPUs far exceeds supply, potentially putting a bottleneck on the pace of AI innovation—but also potentially inspiring new techniques to make the most of GPU power currently available.

Listing image: Getty Images | Aurich Lawson

Photo of Benj Edwards
Benj Edwards Senior AI Reporter
Benj Edwards was a reporter at Ars Technica covering artificial intelligence and technology history.
75 Comments