December 27, 2024

Nvidia Corporation co-founder and CEO Jensen Huang at the Nvidia GPU Technology Conference (GTC) in San Jose, California, USA, on Tuesday, March 19, 2024.

David Paul Morris | David Paul Morris Bloomberg | Getty Images

NVIDIA A 27% gain in May brought its market capitalization to $2.7 trillion, trailing only Microsoft and apple Become one of the world’s most valuable public companies. The chipmaker reported that its sales tripled year-on-year for the third consecutive quarter as demand for artificial intelligence processors soared.

Mizuho Securities estimates that Nvidia controls 70% to 95% of the market for AI chips used to train and deploy models such as OpenAI’s GPT. The 78% gross margin highlights Nvidia’s pricing power and is a surprisingly high number for a hardware company that must make and ship physical products.

Competitor chip manufacturers Intel and Advanced Micro Devices Gross profit margins in the latest quarter were 41% and 47% respectively.

NVIDIA’s position in the AI ​​chip market has been described by some experts as a moat. Its flagship AI graphics processing units (GPUs) like the H100, combined with the company’s CUDA software, put it so far ahead of the competition that turning to alternatives was almost unthinkable.

Still, Nvidia Chief Executive Jensen Huang, whose net worth has grown from $3 billion to about $90 billion over the past five years, said he was “concerned and worried” about the 31-year-old company losing its edge. He acknowledged at a conference late last year that many powerful competitors were emerging.

“I don’t think people are trying to bankrupt me,” Huang explain November. “I probably know they’re working on it, so that’s different.”

Nvidia has promised to release There are new AI chip architectures every yearrather than launching new software every other year as it has historically done, it can more deeply solidify its chips’ position in artificial intelligence software.

But Nvidia’s GPUs aren’t the only chips capable of running the complex mathematical operations that power generative artificial intelligence. If less powerful chips can do the same job, Huang’s paranoia may be justified.

The shift from training AI models to so-called inference (or deploying models) may also provide companies with an opportunity to replace Nvidia GPUs, especially if they are cheaper to buy and run. Nvidia’s flagship chips cost around $30,000 or more, giving customers plenty of incentive to look for alternatives.

“Nvidia wants to have 100% of the data, but customers won’t like Nvidia to have 100% of the data,” said Sid Sheth, co-founder of ambitious rival D-Matrix. “It’s such a big opportunity. It would be unhealthy for any one company to seize it all.”

Founded in 2019, D-Matrix plans to release a semiconductor card for servers later this year, aiming to reduce the cost and latency of running artificial intelligence models. company Upregulate It was $110 million in September.

In addition to D-Matrix, companies ranging from multinationals to emerging startups are vying for a share of the artificial intelligence chip market, which could reach $400 billion in annual sales over the next five years, according to market analysts and AMD. Nvidia generated about $80 billion in revenue over the past four quarters, and Bank of America estimated the company’s AI chip sales last year at $34.5 billion.

Many companies adopting Nvidia GPUs believe that different architectures or certain trade-offs can produce better chips for specific tasks. Device makers are also developing technology that could eventually do the vast amount of AI computing now done on large GPU-based clusters in the cloud.

“No one can deny that today Nvidia is the hardware you want to train and run AI models on,” said Fernando Vidal, co-founder of 3Fourteen Research. told CNBC. “But there has been incremental progress towards leveling the playing field, from very large companies developing their own chips, to smaller startups designing their own chips.”

AMD CEO Lisa Su wants investors to believe there’s enough room in the space for many successful companies.

“The key is there are a lot of options out there,” Su told reporters in December, when her company unveiled its latest artificial intelligence chip. “I think we’re going to see a situation where there’s not just one solution but multiple solutions.”

Other large chip manufacturers

Lisa Su demonstrated the AMD Instinct MI300 chip during her keynote speech at CES 2023 in Las Vegas, Nevada on January 4, 2023.

David Becker | Getty Images

NVIDIA’s top customers

How AWS is designing its own chips to help catch up with Microsoft and Google in the race to generate artificial intelligence

One potential challenge for Nvidia is that it is competing with some of its largest customers. Cloud providers include Google, Microsoft and Amazon Both are building processors for internal use. The Big Three of tech giants, plus Oracleaccounting for more than 40% of Nvidia’s revenue.

Amazon launched its own artificial intelligence chip in 2018, under the brand name Inferentia. Inferentia is now available in its second edition. In 2021, Amazon Web Services debuted Traium for training. Customers can’t buy the chips, but can rent systems through AWS, which claims the chips are more cost-effective than Nvidia’s chips.

Google is probably the cloud provider most committed to its own chips. Since 2015, the company has been using so-called tensor processing units (TPUs) to train and deploy artificial intelligence models. In May, Google released the sixth version of its chip, Trillium, which the company said was used to develop its models, including Gemini and Imagen.

Google also uses Nvidia chips and delivers them through its cloud.

Microsoft hasn’t gone that far yet. company said last year The company is building its own artificial intelligence accelerators and processors, called Maia and Cobalt.

Yuan It’s not a cloud provider, but the company needs a lot of computing power to run its software and website and serve ads. Although Facebook’s parent company is buying billions of dollars worth of Nvidia processors, the company said in April that some of its domestic chips were already deployed in data centers and were “more efficient” than GPUs.

JPMorgan analysts estimated in May that the market for building customized chips for large cloud providers could be worth as much as $30 billion, with potential annual growth of 20%.

Start-up company

Cerebras’ WSE-3 chip is one example of a new class of chips from an upstart designed to run and train artificial intelligence.

brain systems inc.

Venture capitalists see an opportunity for emerging companies to join the game. They invested $6 billion in AI semiconductor companies in 2023, up slightly from $5.7 billion a year ago, according to PitchBook data.

This is a tough area for startups because semiconductors are expensive to design, develop and manufacture. But there are also opportunities for differentiation.

For Silicon Valley artificial intelligence chip maker Cerebras Systems, the focus is on the basic operations and bottlenecks of artificial intelligence, rather than the more general nature of GPUs. The company was founded in 2015 and was valued at $4 billion in its latest financing, according to Bloomberg.

The Cerebras chip WSE-2 combines GPU functionality along with central processing and additional memory into a single device, which is better suited for training large models, CEO Andrew Feldman said.

“We use one giant wafer and they use many small wafers,” Feldman said. “They have the challenge of moving data and we don’t.”

Feldman said his company, part of the Mayo Clinic GSKAnd the U.S. military is winning business as a customer for its supercomputing systems, even competing with Nvidia.

“There’s a lot of competition, and I think that’s healthy for the ecosystem,” Feldman said.

D-Matrix’s Sheth said his company plans to release a card later this year with a small chip that will allow more calculations to be done in memory rather than on a chip such as a GPU. D-Matrix’s product plugs into AI servers alongside existing GPUs, but it offloads work from Nvidia chips and helps reduce the cost of generative AI.

Customers are “very receptive and very willing to bring new solutions to market,” Sheth said.

Apple and Qualcomm

On September 22, 2023, Apple iPhone 15 series devices were launched at The Grove Apple retail store in Los Angeles, California.

Patrick T. Fallon AFP | Getty Images

Don’t miss these exclusive reports from CNBC PRO

Nvidia and Microsoft's performance masks weakness in tech stocks, says Solus' Dan Greenhaus

About The Author

Leave a Reply

Your email address will not be published. Required fields are marked *