The relentless stream of AI headlines makes it hard to get a sense of which businesses are real winners in the AI boom—and which will win in the longer run. To help answer this question The Economist has looked where value has accrued so far and how this tallies with the expected sales of products and services in the AI “stack”, as technologists call the various layers of hardware and software on which AI relies to work its magic. On March 18th many companies up and down the stack will descend on San Jose for a four-day jamboree hosted by Nvidia. With talks on everything from robotics to drug discovery, the shindig will show off the latest AI innovations. It will also highlight furious competition between firms within layers of the stack and, increasingly, between them.
Our analysis examined four of these layers and the companies that inhabit them: AI-powered applications sold to businesses outside the stack; the AI models themselves, such as GPT-4, the brain behind ChatGPT, and repositories of them (for example, Hugging Face); the cloud-computing platforms which host many of these models and some of the applications (Amazon Web Services, Google Cloud Platform, Microsoft Azure); and the hardware, such as semiconductors (made by firms such as AMD, Intel and Nvidia), servers (Dell) and networking gear (Arista), responsible for the clouds’ computing oomph (see chart 1).
Technological breakthroughs tend to elevate new tech giants. The PC boom in the 1980s and 1990s propelled Microsoft, which made the Windows operating system, and Intel, which manufactured the chips needed to run it, to the top of the corporate pecking order. By the 2000s “Wintel” was capturing four-fifths of the operating profits from the PC industry, according to Jefferies, an investment bank. The smartphone era did the same to Apple. Only a few years after it launched the iPhone in 2007, it was capturing more than half of handset-makers’ global operating profits.
The world is still in the early days of the generative-AI epoch. Even so, it has already been immensely lucrative. All told, the 100 or so companies that we examined have together created $8trn in value for their owners since its start—which, for the purposes of this article, we define as October 2022, just before the launch of ChatGPT (see chart 2). Not all of these gains are the result of the AI frenzy—stockmarkets have been on a broader tear of late—but many are.
At every layer of the stack, value is becoming more concentrated in a handful of leading firms. In hardware, model-making and applications, the biggest three companies have increased their share of overall value created by a median of 14 percentage points in the past year and a half. In the cloud layer Microsoft, which has a partnership with ChatGPT’s maker, OpenAI, has pulled ahead of Amazon and Alphabet (Google’s parent company). Its market capitalisation now accounts for 46% of the cloud trio’s total, up from 41% before the release of ChatGPT.
Skimming the cream
The spread of value is uneven between layers, too. In absolute terms the most riches have accrued to the hardware-makers. This bucket includes chip firms (such as Nvidia), companies that build servers (Dell) and those that make networking equipment (Arista). In October 2022 the 27 public hardware companies in our sample were worth around $1.5trn. Today that figure is $5trn. This is what you would expect in a technology boom: the underlying physical infrastructure needs to be built first in order for software to be offered. In the late 1990s, as the internet boom was getting going, providers of things like modems and other telecoms gubbins, such as Cisco and WorldCom, were the early winners.
So far the host of the San Jose gabfest is by far the biggest victor. Nvidia accounts for some 57% of the increase in the market capitalisation of our hardware firms. The company makes more than 80% of all AI chips, according to IDC, a research firm. It also enjoys a near-monopoly in the networking equipment used to yoke the chips together inside the AI servers in data centres. Revenues from Nvidia’s data-centre business more than tripled in the 12 months to the end of January, compared with the year before. Its gross margins grew from 59% to 74%.
Nivdia’ chipmaking rivals want a piece of these riches. Established ones, such as AMD and Intel, are launching rival products. So are startups like Groq, which makes super-fast AI chips, and Cerebras, which makes super-sized ones. Nvidia’s biggest customers, the three cloud giants, are all designing their own chips, too—as a way both to reduce reliance on one provider and to steal some of Nvidia’s juicy margins for themselves. Lisa Su, chief executive of AMD, has forecast that revenues from the sale of AI chips could balloon to $400bn by 2027, from $45bn in 2023. That would be far too much for Nvidia alone to digest.
As AI applications become more widespread, a growing share of that demand will also shift from chips required for training models, which consists in analysing mountains of data in order to teach algorithms to predict the next word or pixel in a sequence, to those needed actually to use them to respond to queries, (“inference”, in tech-speak). In the past year about two-fifths of Nvidia’s AI revenues came from customers using its chips for inference. Experts expect some inference to start moving from specialist graphics-processing units (GPUs), which are Nvidia’s forte, to general-purpose central processing units (CPUs) like those used in laptops and smartphones, which are dominated by AMD and Intel. Before long even some training may be done on CPUs rather than GPUs.
Still, Nvidia’s grip on the hardware market seems secure for the next few years. Startups with no track record will struggle to convince big clients to reconfigure corporate hardware systems for their novel technology. The cloud giants’ deployment of their own chips is still limited. And Nvidia has CUDA, a software platform which allows customers to tailor chips to their needs. It is popular with programmers and makes it hard for customers to switch to rival semiconductors, which CUDA does not support.
Whereas hardware wins the value-accrual race hands down in absolute terms, it is the independent model-makers that have enjoyed the biggest proportional gains. The collective value of 11 such firms we have looked at has jumped from $29bn to about $138bn in the past 16 months. OpenAI is thought to be worth some $100bn, up from $20bn in October 2022. Anthropic’s valuation surged from $3.4bn in April 2022 to $18bn. Mistral, a French startup founded less than a year ago, is now worth around $2bn.
Some of that value is tied up in hardware. The startups buy piles of chips, mostly from Nvidia, in order to train their models. Imbue, which like OpenAI and Anthropic is based in San Francisco, has 10,000 such chips. Cohere, a Canadian rival, has 16,000. These semiconductors can sell for tens of thousands of dollars apiece. As the models get ever more sophisticated, they need ever more. GPT-4 reportedly cost about $100m to train. Some suspect that training its successor could cost OpenAI ten times as much.
Yet the model-makers’ true worth lies in their intellectual property, and the profits that it may generate. The true extent of those profits will depend on just how fierce competition among model providers will get—and how long it will last. Right now the rivalry is white hot, which may explain why the layer has not gained as much value in absolute terms.
Although OpenAI seized an early lead, challengers have been catching up fast. They have been able to tap the same data as the maker of ChatGPT (which is to say text and images on the internet) and, also like it, free of charge. Anthropic’s Claude 3 is snapping at GPT-4’s heels. Four months after the release of GPT-4, Meta, Facebook’s parent company, released Llama 2, a powerful rival that, in contrast to OpenAI’s and Anthropic’s proprietary models, is open and can be tinkered with at will by others. In February Mistral, which has fewer than 40 employees, wowed the industry by releasing an open model whose performance almost rivals that of GPT-4, despite requiring much less computational power to train and run.
Even smaller models increasingly offer good performance at a low price, points out Stephanie Zhan of Sequoia, a venture-capital firm. Some are designed for specific tasks. A startup called Nixtla developed TimeGPT, a model for financial forecasting. Another, Hippocratic AI, has trained its model on data from exams to enter medical school, to give accurate medical advice.
The abundance of models has also enabled the growth of the application layer. The value of the 19 publicly traded software companies in our application group has jumped by $1.1trn, or 35%, since October 2022. This includes big software providers that are adding generative AI to their services. Zoom uses the technology to let users summarise video calls. ServiceNow, which provides tech, human-resources and other support to companies, has introduced chatbots to help resolve customers’ IT queries. Adobe, maker of Photoshop, has an app called Firefly, which uses AI to edit pictures.
Newcomers are adding more variety. “There’s An AI For That”, a website, counts over 12,000 applications, up from fewer than 1,000 in 2022. DeepScribe helps transcribe doctors’ notes. Harvey AI assists lawyers. More idiosyncratically, 32 chatbots promise “sarcastic conversation” and 20 generate tattoo designs. Fierce competition and low barriers to entry mean that some, if not many, applications could struggle to capture value.
Then there is the cloud layer. The combined market capitialsation of Alphabet, Amazon and Microsoft has jumped by $2.5trn since the start of the AI boom. Counted in dollars that is less than three-quarters of the growth of the hardware layer, and barely a quarter in percentage terms. Yet compared with actual revenues that AI is expected to generate for the big-tech trio in the near term, this value creation far exceeds that in all the other layers. It is 120 times the $20bn in revenue that generative AI is forecast to add to the cloud giants’ sales in 2024. The comparable ratio is about 40 for the hardware firms and around 30 for the model-makers.
This implies that investors believe that the cloud giants will be the biggest winners in the long run. The companies’ ratio of share price to earnings, another gauge of expected future profits, tells a similar story. The big three cloud firms average 29. That is above 50% higher than for the typical non-tech firm in the S&P 500 index of large American companies—and up from 21 in early 2023 (see chart 3).
Investors’ cloud bullishness can be explained by three factors. First, the tech titans possess all the ingredients to develop world-beating AI systems: troves of data, armies of researchers, huge data centres and plenty of spare cash. Second, the buyers of AI services, such as big corporations, prefer to do business with established commercial parters than with untested upstarts (see chart 4). Third, and most important, big tech has the greatest potential to control every layer of the stack, from chips to applications. Besides designing some of their own chips, Amazon, Google and Microsoft are investing in both models and applications. Of the 11 model-makers in our sample, nine have the support of at least one of the three giants. That includes the Microsoft-backed OpenAI, Anthropic (Google and Amazon) and Mistral (Microsoft again).
Have the layer cake and eat it
The potential profits that come from controlling more of the layers are also leading hitherto layer-specific firms to branch out. OpenAI’s in-house venture-capital arm has invested in 14 companies since its launch in January 2021, including Harvey AI and Ambience Healthcare, another medical startup. Sam Altman, boss of OpenAI, is reportedly seeking investors to bankroll a pharaonic $7trn chipmaking venture.
Nvidia is becoming more ambitious, too. It has taken stakes in seven of the model-makers, and now offers its own AI models. It has also invested in startups such as Together AI and CoreWeave, which compete with its big cloud customers. At its San Jose event it is expected to unveil a snazzy new GPU and, just maybe, AI tools from other layers of the stack. The AI boom’s biggest single value-creator is in no mood to cede its crown.
© 2023, The Economist Newspaper Limited. All rights reserved.
From The Economist, published under licence. The original content can be found on www.economist.com