Energy Addiction: AI's Next Big Challenge


AllianceBernstein

There’s a big buzz around artificial intelligence (AI) and its potential to change the world. But much less has been said about its energy footprint. Companies that help solve this energy conundrum could enable a sustainable future for this burgeoning technology—and create opportunities for equity investors.

What’s known as “generative” AI uses machine learning to generate content—including text, audio, video and images. OpenAI’s wildly popular ChatGPT is perhaps the most well-known example. There are countless applications for generative AI, from academic writing to audio and video editing to scientific research. Companies everywhere are hunting for AI applications that can enhance productivity and create business benefits in industries ranging from healthcare to investment management.

But here’s the rub: AI requires massive computational power to train models. And that raises a thorny issue—namely, the energy impact of AI.

Generative AI Is an Energy Hog

What’s behind the magic of machine learning? There are two primary stages. The first is training, which involves gathering information so that machines can learn everything possible to create a model. The second is inference, whereby the machine uses that model to generate content, analyse new data and produce actionable results.

All of this requires energy. The more powerful and complex the AI model, the greater the training time and energy required (Display). 

AI Models’ Computational Complexity Requires Plenty of Power

Training time for AI models in petaFLOP/s-days*

Historical analysis does not guarantee future results.
*FLOPs (floating-point operations per second) is a measure of compute performance used in deep-learning models that require floating-point operations. PetaFLOP/s-days represents the number of days required to train a particular model, assuming that machine training the model performed a fixed amount of computation (1,015 neural net operations per second for one entire day). 
As of June 30, 2023.  
Source: Bank of America, NVIDIA and AllianceBernstein (AB)

OpenAI’s GPT-3 model is illustrative. The energy needed to train GPT-3 could power an average American’s home for more than 120 years, according to a report from Stanford University. Meantime, Bay Area chipmaker NVIDIA notes that energy requirements for training models that include transformers—a form of deep-learning architecture—have increased by 275 times every two years.

The Many Sources of Energy Consumption

AI’s energy consumption will come from many corners. In addition to training and running large models, the proliferation of AI-assisted products, including AI search and chatbots, will gobble up terawatts.

Increasingly complex models will, in turn, require the use of more specialized hardware, such as graphic processing units (GPUs). The good news is that GPUs deliver much more performance-per-watt than traditional central processing units (CPUs), which could offset the overall power requirements to train and run AI models.

Ultimately, these drivers of energy consumption will accelerate the construction of power-hungry data centres, which already account for nearly 1% of global energy use, according to the International Energy Agency. Even before AI began to take off, studies predicted a sharp increase in data centre construction, driven by the energy needs of new technologies.

There’s also the issue of emissions to consider. In particular, investors are pushing companies to measure Scope 3 emissions—upstream and downstream emissions that can be difficult to quantify. As AI use increases, the Scope 3 emissions of all data users—including firms that traditionally have low carbon footprints—are likely to grow correspondingly.

How Are Companies Addressing the AI Energy Conundrum?

Fortunately, companies are beginning to address the enormous AI energy challenge. These include firms that are central to AI and those only nibbling at the periphery. We think investors should pay attention to three key areas:

Hardware and Software: Reducing AI-related energy use will require new processor architectures. US semiconductor makers like AMD and NVIDIA are focused on delivering more energy-efficient performance. In fact, AMD has set a goal of increasing the energy efficiency of its processors and accelerators used in AI training and high-performance computing by 30 times over a five-year period. According to NVIDIA, its GPU-based servers in some applications, such as large-language model training, use 25 times less energy than CPU-based alternatives. As GPUs from AMD, NVIDIA and others take share from CPUs in data centres, energy efficiency should increase even further.

Conserving energy will also require advanced transistor-packaging techniques. Technologies such as dynamic voltage frequency scaling and thermal management will be required to produce more efficient machine learning. We believe companies involved in semiconductor chip production and inspection, including Taiwanese chipmaker TSMC and Netherlands-based ASML, will have a significant role to play in bringing these new innovations to market.

Investors will also be hearing more about power semiconductors, which help improve the power management of AI servers and data centres. Power semiconductors regulate current and can lower overall energy use by integrating more functionality in smaller footprints. Firms like Kirkland, Washington-based Monolithic Power Systems and German semiconductor manufacturer Infineon Technologies are at the forefront of their development, in our view.

Improvements in Data Centre Design: As AI adoption fuels the expansion of data centre capacity, firms that supply data centre components could reap benefits. Key components include power supplies, optical networking, memory systems and cabling. Tech companies that use the data centres themselves—think Amazon.com, Google and Microsoft—also have a strong incentive to continue improving data centre design and energy consumption.

Coming full circle, AI itself is being used to optimize data centre operations. In 2022, Google DeepMind released the results of a three-month experiment that involved training a learning agent called BCOOLER to optimize Google’s data centre cooling procedures. The result: BCOOLER achieved roughly 13% energy savings—underscoring that energy efficiency is improving in data centres, even as their numbers grow.

Renewable Energy: Renewables made up 21.5% of US electricity generation in 2022, according to the Energy Information Administration. With 80% of the US power grid non-renewable, near-term power could come from traditional fossil fuels.

But over time, AI demand could open the door for more renewable energy use. That’s especially true given that AI data centres will be operated by the likes of Microsoft and Google’s parent company, Alphabet, Inc., whose net zero policies are among the industry’s best. As a result, we expect that accelerated adoption of AI could improve the investment prospects of the entire renewable-power ecosystem.

Investing in Energy Solutions

In all these areas, we believe that investors should search for quality companies with a technological advantage, persistent pricing power, healthy free-cash-flow generation and resilient business models. Companies with strong fundamentals that are poised to participate in and benefit from increased demand for energy-efficient AI capabilities could provide attractive opportunities for equity investors with a sustainable focus and those with an absolute-return mandate.

As AI adoption accelerates and search engines are replaced by chatbots, the energy impact of this revolutionary form of machine learning should not be overlooked. Initiatives aimed at creating a more energy-efficient AI ecosystem might not be in the spotlight now, but they could eventually unlock attractive return potential for investors who can spot the potential solutions early.

 

Any information provided by AllianceBernstein Australia Limited (“ABAL”)—ABN 53 095 022 718, AFSL 230 698. Information in this presentation is only intended for persons that qualify as “wholesale clients,” as defined by the Corporations Act 2001, and is not to be construed as advice. The views expressed may not represent the views of ABAL.  This information is provided solely for informational purposes and is not an offer to buy or sell securities or to provide any service. The information, forecasts and opinions set forth in this presentation reflects our most recent opinion and which may change, have not been prepared for any recipient’s specific investment objectives, financial situation or particular needs. Neither this presentation nor the information contained in it are intended to take the place of professional advice. You should not take action on specific issues based on the information contained in this presentation without first obtaining professional advice.