《麥肯錫:2025計算成本:AI 驅動下數據中心基礎設施萬億級投資競賽報告(英文版)(12頁).pdf》由會員分享,可在線閱讀,更多相關《麥肯錫:2025計算成本:AI 驅動下數據中心基礎設施萬億級投資競賽報告(英文版)(12頁).pdf(12頁珍藏版)》請在三個皮匠報告上搜索。
1、Technology,Media&Telecommunications PracticeThe cost of compute:A$7 trillion race to scale data centersAI is fueling high demand for compute power,spurring companies to invest billions of dollars in infrastructure.But with future demand uncertain,investors will need to make calculated decisions.This
2、 article is a collaborative effort by Jesse Noffsinger,Mark Patel,and Pankaj Sachdeva,with Arjita Bhan,Haley Chang,and Maria Goodpaster,representing views from McKinseys Technology,Media&Telecommunications Practice.April 2025Amid the AI boom,compute power is emerging as one of this decades most crit
3、ical resources.In data centers across the globe,millions of servers run 24/7 to process the foundation models and machine learning applications that underpin AI.The hardware,processors,memory,storage,and energy needed to operate these data centers are collectively known as compute powerand there is
4、an unquenchable need for more.Our research shows that by 2030,data centers are projected to require$6.7 trillion worldwide to keep pace with the demand for compute power.Data centers equipped to handle AI processing loads are projected to require$5.2 trillion in capital expenditures,while those powe
5、ring traditional IT applications are projected to require$1.5 trillion in capital expenditures(see sidebar“What about non-AI workloads?”).Overall,thats nearly$7 trillion in capital outlays needed by 2030a staggering number by any measure.1What about non-AI workloads?While AI workloads dominate the c
6、onversation,non-AI processing loads remain a significant portion of data center activity.These include traditional enterprise IT tasks such as web hosting,enterprise resource planning systems,email,and file storage.Non-AI loads are less compute-intensive and can operate efficiently on central proces
7、sing units rather than the specialized graphics processing units or AI accelerators that AI workloads require.They also tend to have more predictable usage patterns and lower power densities,which allow for less demanding cooling and energy requirements.As a result,data centers focused on non-AI pro
8、cessing typically have different infrastructure needs,capital intensity,and operational considerations compared with those optimized for AI.To meet this demand,companies across the compute power value chain must strike a balance between deploying capital quickly and doing so prudently.To improve the
9、 odds that their data center investments will provide strong returns,companies can tackle projects in stages,assessing ROI at each step.Still,a lack of clarity about future demand makes precise investment calculations difficult.The compute power value chain is complexfrom the real estate developers
10、that build data centers to the utilities that power them,to the semiconductor firms that produce chips to the cloud service hyperscalers that host trillions of terabytes of data.Leaders across this value chain know that they must invest in compute power to accelerate AI growth.But their challenge 1
11、McKinsey Data Center Demand Model,McKinsey Data Center Capex TAM Model,and expert interviews.The cost of compute:A$7 trillion race to scale data centers2is formidable:deciding how much capital to allocate to which projects,all while remaining uncertain of how AIs future growth and development will i
12、mpact compute power demand.Will hyperscalers continue shouldering the cost burden,or will enterprises,governments,and financial institutions step in with new financing models?Will demand for data centers rise amid a continued surge in AI usage,or will it fall as technological advances make AI less c
13、ompute-heavy?One thing is certain:The stakes are high.Overinvesting in data center infrastructure risks stranding assets,while underinvesting means falling behind.This article,based on McKinsey research and analysis,provides companies across the compute power value chain with an overview of the inve
14、stment landscape for the next five years.Despite the rigor behind these forecasts,we acknowledge that AI is a radically evolving space.Our analysis is built on thoroughly researched hypotheses,but there are critical uncertainties that cannot yet be quantified.Predicting the compute power demand curv
15、eTo decide how much to invest in compute power,companies should first accurately forecast future demanda challenging task given that the AI sector is shifting so rapidly.Our research shows that global demand for data center capacity could almost triple by 2030,with about 70 percent of that demand co
16、ming from AI workloads(Exhibit 1).However,this projection hinges on two key uncertainties:AI use cases.The value in AI lies at the application layerhow enterprises turn AI into real business impact.If companies fail to create meaningful value from AI,demand for compute power could fall short of expe
17、ctations.Conversely,transformative AI applications could fuel even greater demand than current projections suggest.Rapid innovation cycles and disruptions.Continuous advancements in AI technologies,such as processors,large language model(LLM)architectures,and power consumption,could significantly en
18、hance efficiency.For instance,in February 2025,Chinese LLM player DeepSeek reported that its V3 model achieved substantial improvements in training and reasoning efficiency,notably reducing training costs by approximately 18 times and inferencing costs by about 36 times,compared with GPT-4o.2 Howeve
19、r,preliminary analysis suggests that these types of efficiency gains will likely be offset by increased experimentation and training across the broader AI market.As a result,efficiency gains may not substantially impact overall compute power demand over the long term.3AI demand alone will require$5.
20、2 trillion in investmentWe calculate that companies across the compute power value chain will need to invest$5.2 trillion into data centers by 2030 to meet worldwide demand for AI alone.We based this figure on extensive analysis and key assumptions,including a projected 156 gigawatts(GW)of AI-2 Mani
21、sh Singh,“DeepSeek punctures AI leaders spending plans,and what analysts are saying,”TechCrunch,January 27,2025;Wayne Williams,“OpenAI spent$80M to$100M training GPT-4;Chinese firm claims it trained its rival AI model for$3 million using just 2,000 GPUs,”TechRadar,December 2,2024;“Independent analys
22、is of AI models and API providers,”Artificial Analysis,January 27,2025.3 This aligns with the concept of Jevons Paradox,which posits that improvements in efficiency can lead to increased overall demand.In the context of AI,more efficient and accessible computing resources could spur greater adoption
23、 and utilization,potentially offsetting the anticipated reductions in compute demand.The cost of compute:A$7 trillion race to scale data centers3Exhibit 1related data center capacity demand by 2030,with 125 incremental GW added between 2025 and 2030.This$5.2 trillion figure reflects the sheer scale
24、of investment required to meet the growing demand for AI compute powera significant capital commitment that underscores the magnitude of the challenge ahead(see sidebar“The scale of investment”).Amid the uncertainty about future needs for compute power,we created three investment scenarios ranging f
25、rom constrained to accelerated demand(Exhibit 2).In the first of our three scenarios,growth accelerates significantly and 205 incremental GW of AI-related data center capacity is added between 2025 and 2030.This would require an estimated$7.9 trillion in capital expenditures.The second scenario is t
26、he one we use in this article:Demand grows,but not as much as in the first scenario,and the expected capital expenditure is$5.2 trillion.In our third scenario,in which demand is more constrained,with 78 incremental GW added in the next five years,the total capital expenditure is$3.7 trillion(see sid
27、ebar“Methodology”).The cost of compute:A$7 trillion race to scale data centers4The scale of investmentTo put the trillion-dollar size of investment needed by 2030 into perspective,consider these unrelated statistics that illustrate the sheer scale of capital needed:Labor.$500 billion in labor costs
28、is roughly equivalent to 12 billion labor hours(six million people working full time for an entire year).1Fiber.$150 billion worth of fiber is equivalent to installing three million miles of fiber-optic cablesenough to circle the Earth 120 times.2Power generation.$300 billion worth of power generati
29、on is equivalent to adding 150 to 200 gigawatts of gas,which would be enough to power 150 million homes for a yearmore than the total number of households in the United States.31 Estimated conservatively,using a high-end hourly wage of$40(for roles ranging from construction workers to data center te
30、chnicians),assuming a standard 40-hour workweek and 52 weeks per year.“Occupational employment and wage statistics survey by occupationMay 2024,”US Bureau of Labor Statistics news release,April 2,2025.2 Estimated based on an average of$50,000 as the per-mile aerial installation cost of fiber-optic c
31、ables.Jonathan Kim,“Fiber optic network construction:Process and build costs,”Dgtl Infra,January 15,2024.3 Estimated assuming full utilization and average US household electricity consumption of 10,500 kilowatt-hours per year.Based on data from the US Energy Information Administration.MethodologyCap
32、ital expenditure estimates in this article are derived from McKinseys proprietary data center demand model,which projects data center capacity under multiple scenarios shaped by factors such as semiconductor supply constraints,enterprise AI adoption,efficiency improvements,and regulatory challenges.
33、Investment requirements were calculated by translating demand projections for gigawatt capacity into capital expenditures across major cost categories,including power(for example,generation,transmission),data center infrastructure(for example,electrical,mechanical,site,shell),and IT equipment(for ex
34、ample,AI accelerators,networking,storage).The cost of compute:A$7 trillion race to scale data centers5Exhibit 2In any scenario,these are staggering investment numbers.They are fueled by several factors:Mass adoption of gen AI.The foundation models that underpin gen AI require significant compute pow
35、er resources to train and operate.Both training and inference workloads are contributing to infrastructure growth,with inference expected to become the dominant workload by 2030.Enterprise integration.Deploying AI-powered applications across industriesfrom automotive to financial servicesdemands mas
36、sive cloud computing power.As use cases grow,AI applications will grow more sophisticated,integrating specialized foundation models tailored to specific domains.Competitive infrastructure race.Hyperscalers and enterprises are racing to build proprietary AI capacity to gain competitive advantage,whic
37、h is fueling the construction of more and more data centers.These“builders”(as further described below)hope to gain competitive advantage by achieving scale,optimizing across data center tech stacks,and ultimately driving down the cost of compute.The cost of compute:A$7 trillion race to scale data c
38、enters6Geopolitical priorities.Governments are investing heavily in AI infrastructure to enhance security,economic leadership,and technological independence.Where is the investment going?To qualify our$5.2 trillion investment forecast for AI infrastructure,its important to note that our analysis lik
39、ely undercounts the total capital investment needed,as our estimate quantifies capital investment for only three out of five compute power investor archetypesbuilders,energizers,and technology developers and designersthat directly finance the infrastructure and foundational technologies necessary fo
40、r AI growth(see sidebar“Five types of data center investors”).Approximately 15 percent($0.8 trillion)of investment will flow to builders for land,materials,and site development.Another 25 percent($1.3 trillion)will be allocated to energizers for power generation and transmission,cooling,and electric
41、al equipment.The largest share of investment,60 percent($3.1 trillion),will go to technology developers and designers,which produce chips and computing hardware for data centers.The other two investor archetypes,operators,such as hyperscalers and colocation providers,and AI architects,which build AI
42、 models and applications,also invest in compute power,particularly in areas such as AI-driven automation and data center software.But quantifying their compute power investment is challenging because it overlaps with their broader R&D spending.Despite these projected capital requirements,our researc
43、h shows that current investment levels lag demand.In dozens of client interviews,we found that CEOs are hesitant to invest in compute power capacity at maximum levels because they have limited visibility into future demand.Uncertainty about whether AI adoption will continue its rapid ascent and the
44、fact that infrastructure projects have long lead times make it difficult for companies to make informed investment decisions.Many companies are unsure whether large capital expenditures on AI infrastructure today will produce measurable ROI in the future.So how can business leaders move forward conf
45、idently with their investments?As a first step,they can determine where their organizations fall within the compute power ecosystem.Five archetypes of AI infrastructure investorsWho are the investors behind the multitrillion-dollar race to fund AI compute power?We have identified five key investor a
46、rchetypes,each navigating distinct challenges and opportunities,and detailed how much they could spend in the next five years.1.BuildersWho they are:real estate developers,design firms,and construction companies expandingdata center capacityThe cost of compute:A$7 trillion race to scale data centers
47、7Five types of data center investorsAs AI drives a surge in compute power demand,five types1 of organizations are leading the massive capital investments required to scale data centers:Builders:real estate developers,design firms,and construction companies that expand and upgrade data centers,such a
48、s Turner Construction and AECOMEnergizers:companies that supply the electricity and cooling systems essential for data center operations,including utilities like Duke Energy and Entergy and infrastructure and equipment providers like Schneider Electric and VertivTechnology developers and designers:s
49、emiconductor companies that develop the chips powering AI workloads,such as NVIDIA and Intel,and computing hardware suppliers such as Foxconn and FlexOperators:cloud providers and co-location firms that own and run large-scale data centers,such as Amazon Web Services,Google Cloud,and EquinixAI archi
50、tects:companies developing AI models and infrastructure,including OpenAI and Anthropic1 The companies listed as examples in each archetype category may also span adjacent categories.For instance,hyperscalers in the operators category(Amazon Web Services,Google Cloud)are developing specialized comput
51、ing hardware and investing in both their own and third-party LLM productsactivities that align with the roles of technology developers and designers and AI architects,respectively.AI workload capital expenditure:$800 billionNon-AI workload capital expenditure:$100 billionKey investments:land and mat
52、erial acquisition,skilled labor,site developmentOpportunities.Builders that optimize site selection can secure prime locations,reduce construction timelines,and integrate operational feedback early,ensuring faster deployment and higher data center efficiency.Challenges.Labor shortages could impact t
53、echnician and construction worker availability,while location constraints could limit site selection options.Meanwhile,increased rack power density could create space and cooling challenges.The cost of compute:A$7 trillion race to scale data centers8Solutions.Forward-thinking builders can find solut
54、ions to core challenges,adding certainty to their investment decisions.For example,some are solving the labor shortage issue by adopting modular designs that streamline the construction process,such as off-site construction of large components that can be assembled on-site.2.EnergizersWho they are:u
55、tilities,energy providers,cooling/electrical equipment manufacturers,andtelecom operators building the power and connectivity infrastructure for AI data centers4AI workload capital expenditure:$1.3 trillionNon-AI workload capital expenditure:$200 billionKey investments:power generation(plants,transm
56、ission lines),cooling solutions(air cooling,direct-to-chip liquid cooling,immersion cooling),electrical infrastructure(transformers,generators),network connectivity(fiber,cable)Opportunities.Energizers that scale power infrastructure and innovate in sustainable energy solutions will be best position
57、ed to benefit from hyperscalers growing energy demands.Challenges.Powering data centers could stall due to existing grid weaknesses and solving heat management challenges from rising processor densities remains an obstacle.Energizers also face clean-energy transition requirements and lengthy grid co
58、nnection approval processes.Solutions.With over$1 trillion in investment at stake,energizers are finding ways to deliver reliable power while driving ROI.They are making substantial investments in emerging power-generation technologiesincluding nuclear,geothermal,carbon capture and storage,and long-
59、duration energy storage.They are also doubling down on efforts to bring as much capacity online as quickly as possible across both renewable sources and traditional energy infrastructure,such as gas and fossil fuels.What is changing now is the sheer scale of that demand,which brings a new urgency to
60、 build power capacity at unprecedented speed.As demandespecially for clean energysurges,power generation is expected to grow rapidly,with renewables projected to account for approximately 45 to 50 percent of the energy mix by 2030,up from about a third today.53.Technology developers and designersWho
61、 they are:semiconductor firms and IT suppliers producing chips and computinghardware for data centersAI workload capital expenditure:$3.1 trillionNon-AI workload capital expenditure:$1.1 trillion4 For more on how utility and energy providers are investing in AI infrastructure,see“How data centers an
62、d the energy sector can sate AIs hunger for power,”McKinsey,September 17,2024;for more on how telecom operators are investing in AI infrastructure,see“AI infrastructure:A new growth avenue for telco operators,”McKinsey,February 28,2025.5“Global Energy Perspective,”McKinsey,September 17,2024.The cost
63、 of compute:A$7 trillion race to scale data centers9 Key investments:GPUs,CPUs,memory,servers,and rack hardwareOpportunities.Technology developers and designers that invest in scalable,future-ready technologies supported by clear demand visibility could gain a competitive edge in AI computing.Challe
64、nges.A small number of semiconductor firms control the market supply,stifling competition.Capacity building remains insufficient to meet current demand,while at the same time,shifts in AI model training methods and workloads make it difficult to predict future demand for specific chips.Solutions.Tec
65、hnology developers and designers have the most to gain in the compute power race because they are the ones providing the processors and hardware that do the actual computing.Demand for their products is currently high,but their investment needs are also the greatestmore than$3 trillion over the next
66、 five years.A small number of semiconductor firms have a disproportionate influence on industry supply,making them potential chokepoints in compute power growth.Technology developers and designers can mitigate this risk by expanding fabrication capacity and diversifying supply chains to prevent bott
67、lenecks.4.Operators Who they are:hyperscalers,colocation providers,GPU-as-a-service platforms,andenterprises optimizing their computing resources by improving server utilization and efficiency AI workload capital expenditure:not included in this analysis Non-AI workload capital expenditure:not inclu
68、ded in this analysis Key investments:data center software,AI-driven automation,custom siliconOpportunities.Operators that scale efficiently while balancing ROI,performance,and energy use can drive long-term industry leadership.Challenges.Immature AI-hosted applications can obscure long-term ROI calc
69、ulations.Inefficiencies in data center operations are driving up costs,but uncertainty in AI demand continues to disrupt long-term infrastructure planning and procurement decisions.Solutions.While data centers today operate at high-efficiency levels,the rapid pace of AI innovation will require opera
70、tors to optimize both energy consumption and workload management.Some operators are improving energy efficiency in their data centers by investing in more effective cooling solutions and increasing rack stackability to reduce space requirements without sacrificing processing power,for example.Others
71、 are investing in AI model development itself to create architectures that need less compute power to be trained and operated.The cost of compute:A$7 trillion race to scale data centers105.AI architects Who they are:AI model developers,foundation model providers,and enterprises buildingproprietary A
72、I capabilities AI workload capital expenditure:not included in this analysis Non-AI workload capital expenditure:not included in this analysis Key investments:model training and inference infrastructure,algorithm researchOpportunities.AI architects that develop architectures that balance performance
73、 with lower compute requirements will lead the next wave of AI adoption.Enterprises investing in proprietary AI capabilities can gain competitiveness by developing specialized models tailored to their needs.Challenges.AI governance issues,including bias,security,and regulation,add complexity and can
74、 slow development.Meanwhile,inference poses a major unpredictable cost component,and enterprises are facing difficulties demonstrating clear ROI from AI investments.Solutions.The escalating computational demands of large-scale AI models are driving up the costs to train them,particularly regarding i
75、nference,or the process where trained AI models apply their learned knowledge to new,unseen data to make predictions or decisions.Models with advanced reasoning capabilities,such as OpenAIs o1,require significantly higher inference costs.For example,it costs six times more for inference on OpenAIs o
76、1 compared with the companys nonreasoning GPT-4o.To bring down inference costs,leading AI companies are optimizing their model architectures by using techniques like sparse activations and distillation.These solutions reduce the computational power needed when an AI model generates a response,making
77、 operations more efficient.Critical considerations for AI infrastructure growthAs companies plan their AI infrastructure investments,they will have to navigate a wide range of potential outcomes.In a constrained-demand scenario,AI-related data center capacity could require$3.7 trillion in capital ex
78、penditureslimited by supply chain constraints,technological disruptions,and geopolitical uncertainty.These barriers are mitigated,however,in an accelerated-demand scenario,leading to investments as high as$7.9 trillion.Staying on top of the evolving landscape is critical to making informed,strategic
79、 investment decisions.Some of the uncertainties investors must consider include:Technological disruptions.Breakthroughs in model architectures,including efficiency gains in compute utilization,could reduce expected hardware and energy demand.The cost of compute:A$7 trillion race to scale data center
80、s11Supply chain constraints.Labor shortages,supply chain bottlenecks,and regulatory hurdles could delay grid connections,chip availability,and data center expansionslowing overall AI adoption and innovation.To address supply chain bottlenecks for critical chips,semiconductor companies are investing
81、significant capital to construct new fabrication facilities,but this construction could stall due to regulatory constraints and long lead times from upstream equipment suppliers.Geopolitical tensions.Fluctuating tariffs and technology export controls could introduce uncertainty in compute power dema
82、nd,potentially impacting infrastructure investments and AI growth.The race for competitive advantageThe winners of the AI-driven computing era will be the companies that anticipate compute power demand and invest accordingly.Companies across the compute power value chain that proactively secure crit
83、ical resourcesland,materials,energy capacity,and computing powercould gain a significant competitive edge.To invest with confidence,they can take a three-pronged approach.First,investors will need to understand demand projections amid uncertainty.Companies should assess AI computing needs early,anti
84、cipate potential shifts in demand,and design scalable investment strategies that can adapt as AI models and use cases evolve.Second,investors should find ways to innovate on compute efficiency.To do so,they can prioritize investments in cost-and energy-efficient computing technologies,optimizing per
85、formance while managing power consumption and infrastructure costs.Third,they can build supply-side resilience to sustain AI infrastructure growth without overextending capital.This will require investors to secure critical inputs such as energy and chips,optimize site selection,and build flexibilit
86、y into their supply chains.Striking the right balance between growth and capital efficiency will be critical.Investing strategically is not just a race to scale data infrastructureits a race to shape the future of AI itself.We are celebrating the 60th birthday of the McKinsey Quarterly with a yearlo
87、ng campaign featuring four issues on major themes related to the future of business and society,as well as related interactives,collections from the magazines archives,and more.This article will appear in the fourth themed issue,which will launch in July.Sign up for the McKinsey Quarterly alert list
88、 to be notified as soon as other new Quarterly articles are published.Jesse Noffsinger is a partner in McKinseys Seattle office,where Maria Goodpaster is an associate partner;Mark Patel is a senior partner in the Bay Area office,where Haley Chang is a consultant;Pankaj Sachdeva is a senior partner i
89、n the Philadelphia office;and Arjita Bhan is a knowledge expert in the Boston office.The authors wish to thank Andrea Boza Zanatta,Jason Amri,Rishi Gupta,Senem Bilir,and Shraddha Kumar for their contributions to this article.This article was edited by Kristi Essick,an executive editor in the Bay Area office.Copyright 2025 McKinsey&Company.All rights reserved.The cost of compute:A$7 trillion race to scale data centers12