09/24/2025 | Press release | Distributed by Public on 09/23/2025 21:57
The exponential growth in AI usage, alongside rising industrial energy needs, is fueling an unprecedented demand for power. However, the aging U.S. grid is struggling to deliver the high-density load required by modern industries. Most power transformers, circuit breakers, and transmission lines are over 25 years old, with many exceeding 30 years and wearing down fast, according to the U.S. Department of Energy's infrastructure review. Significant transmission upgrades are needed to support the additional load coming online. The Electric Reliability Council of Texas (ERCOT), which manages the state's power grid, estimates that meeting ever-growing power demand there will require a roughly $30 billion investment to lay thousands of miles of long-distance transmission lines.
The U.S. relies on multiple independent electrical grids, including PJM, MISO, and ERCOT. Each operates under a different regulatory framework, making the development and interconnection of large-load energy campuses complex. As grid operators assess future supply, demand, and transmission constraints, interconnection queues continue to grow, making sites with approved access to large amounts of power rare and highly coveted. In most markets, the wait time to plug into the grid is now four years, according to real estate brokerage JLL.
This report examines in depth the demand projections for data centers, the AI trends driving them, and the strong position of Galaxy's flagship West Texas data center campus, Helios, to power the next generation of AI technology.
Several experts have projected a massive scale-up in demand for power from data centers. First, the U.S. Department of Energy (DOE) projects that data center power consumption will double or triple by 2028, rising from 176 terawatt-hours (TWh) in 2023 to between 325 TWh and 580 TWh in 2028. Accordingly, McKinsey projects the capacity demanded from U.S. data centers to rise from 25gigawatts (GW) in 2024 to more than 80 GW by 2030. For Texas alone, ERCOT projects data center energy demand to increase by 24 GW by 2031, with total load growing to 145 GW. This is a significant increase from ERCOT's 2025 peak power projection of 87 GW.
AI companies - such as hyperscalers, neoclouds, and AI labs - are rapidly leasing and building data center capacity to meet escalating computational needs. Despite skepticism among analysts that data center capex projections would remain high following the DeepSeek AI model's apparent efficiency gains, hyperscalers including Amazon, Oracle, Google, Microsoft, and Meta have reiterated or increased their 2025 capex commitments. Collectively, these firms are expected to spend more than $360 billion, signaling sustained high demand for high-density data center space.
We believe that the monetization potential of AI companies is fundamentally linked to the scale of data center infrastructure supporting their workloads. Greater infrastructure capacity - measured in critical IT load - enables greater computing power, which in turn allows for the development of more advanced AI models and supports more inference workloads, driving substantial revenue growth.
For instance, training state-of-the-art models requires thousands of graphics processing units (GPUs), while large-scale inference allows companies to serve more users, enhancing monetization. This point is further reinforced by NVIDIA CEO Jensen Huang, who holds that the revenue-generating potential of a GPU-hosting "AI factory" is proportional to the number of tokens (units of data) it can generate and the speed at which they are produced (tokens per second). Consequently, access to large-scale power and suitable land for AI and high-performance computing (HPC) data centers will be extremely valuable as AI adoption continues to grow. As McKinsey notes, "If you're looking to add 100, 200, 500 megawatts-or even a gigawatt-of power at a single campus, there aren't that many places where you can just attach that to the grid." The consulting firm's research further emphasizes that "substantial new grid infrastructure will be needed to support the rising AI economy."
On the supply side, data center availability in primary markets has become severely constrained. According to real estate brokerage CBRE, "demand is outpacing supply in nearly every major market, driving elevated lease rates and prompting rapid market expansion," with North America's data center vacancy rate hitting an all-time low of 1.6% as hyperscale and AI occupiers race to secure power and capacity years in advance. As a result, over the past several months, new data center construction has started expanding outside traditional hubs such as Northern Virginia, to more rural areas including West Texas, where there is more power available. We believe ERCOT's grid expansion initiatives and Texas's business-friendly environment will differentiate the state as one of the fastest-growing regions for large-load consumers, positioning Texas to become a leading hub for data centers in the U.S. On April 24, the Public Utility Commission of Texas approved plans to construct the first 765 kilovolt (kV) extra high voltage transmission lines in the ERCOT region to support "the growing electricity needs of West Texas communities."
The Helios campus has an approved capacity of 800MW and an additional 2.7 GW of expansion capacity under study. At a total of 3.5 GW of potential gross capacity, Helios will be one of the largest single-site data center campuses in the world. This substantial capacity presents significant opportunities for AI companies to advance the development of smarter and more capable models. The chart below shows that the total potential critical IT capacity at Helios is larger than those of several primary and secondary data center markets combined.
This section explores how AI scaling laws are driving the demand for power and compute resources across the training and inference stack. Three distinct scaling laws govern AI performance: pre-training scaling, post-training scaling, and test-time scaling (also known as inference-time scaling). These scaling laws, widely validated and agreed upon across the industry, can be individually optimized to boost model performance, unlocking higher levels of intelligence-from perception-based AI and generative models to advanced agentic capabilities.
One of major trends we've been observing in AI is that frontier models are becoming bigger, more computationally intensive, and more demanding of energy. According to the Stanford AI Index Report, the training compute required to train frontier models has doubled every five months, dataset sizes have doubled every eight months, and the power required has doubled annually. These findings are supported by data from Epoch AI, which indicates that the training compute required for frontier models grows at a rate of 4.2x per year, available training data expands at 3.5x per year, and the power needed for training grows at 2.1x per year.
Furthermore, Epoch AI's analysis attributes the 4.2x annual training compute growth in frontier models to three key factors: a 1.7x increase in training hardware quantity (more GPUs), a 1.5x increase in training duration, and a 1.4x improvement in hardware performance. The study concludes that training compute growth is driven by larger clusters, longer training duration, and better hardware performance.
Furthermore, Epoch AI notes that once models are trained beyond 10^24 total floating-point operations (FLOP), performance continues to improve with additional training compute. Across most benchmarks, accuracy rises by roughly 12 percentage points for every 10x increase in compute. On the GPQA Diamond benchmark-which evaluates advanced language models on their ability to answer challenging PhD-level scientific questions in chemistry, physics, and biology-each 10× increase in compute yields about a 14-percentage-point gain in accuracy, further underscoring the importance of scaling pre-training compute.
Meanwhile, the power draw required to train frontier models has been rapidly increasing, with some recent models such as Grok 3 estimated to have required over 100 MW of power capacity to train. This is still a relatively small amount of capacity compared to the size of certain mega-campuses like Helios, showing the enormous potential capabilities that could be unlocked by training multi-hundred MW models.
At Helios, we are excited to play a major role in scaling AI initiatives. As clusters grow larger and more energy-intensive, the Helios data center campus, with reliable access to power at scale, a vast land footprint, ample water supply, and low latency, is ideally equipped to meet the power demands of training next-generation frontier models. Our infrastructure is purpose-built to support billion-dollar clusters, delivering scalability and reliability for cutting-edge AI workloads.
Alongside growth in AI training, several key innovations in the inference stack have emerged that have fueled an increase in demand for compute resources.
Reasoning workloads, unlike traditional LLM prompting, involve complex, recursive inference processes that enable models to solve problems by iteratively analyzing and synthesizing information. These models, often designed for such tasks as logical reasoning, decision-making, or code generation, break down queries into multiple steps, evaluating intermediate results and refining outputs through repeated computations. This recursive nature demands significantly higher computational resources, as the model processes multiple iterations of data, context, and logic, leading to exponentially greater compute and power requirements compared to simpler, single-pass LLM queries. Reasoning models have significantly increased the demand for compute at the inference level, requiring significantly more tokens than traditional large language model (LLM) prompting due to their complex recursive inference processes.
A prime example is the DeepSeek-R1 model, a 671-billion-parameter mixture-of-experts (MoE) reasoning model optimized for math, coding, and logical tasks. DeepSeek-R1 leverages inference-time scaling, also known as test-time scaling or AI reasoning, where additional computational resources are allocated during inference to evaluate multiple potential outcomes. This approach has been shown to produce highly accurate results, particularly in complex tasks including generating optimized GPU attention kernels - a benchmark that evaluates numerical correctness (accuracy) and performance, including speed and resource efficiency, of attention mechanisms. This test enables AI models to optimize for the most relevant parts of query entries when reasoning. Below is a snapshot of the reasoning framework used by DeepSeek-R1 provided by NVIDIA.
The reasoning framework applied to DeepSeek-R1 significantly improved the model's accuracy. As shown in the chart below, the percentage of problems solved follows a logarithmic distribution; performance improves as inference time extends.
Notably, inference-time scaling allows models to allocate more compute during inference, enabling exploration of multiple solutions, which results in optimized kernels but significantly increases compute consumption. This approach has shown the potential to enhance model intelligence, significantly increasing computational demand during inference and amplifying the need for robust, scalable power sources.
DeepSeek's recent breakthroughs in software and cost efficiency improvements have sparked debate over their implications for compute demand and data center growth. When combined with ongoing improvements in hardware performance, these developments have led some in the market to speculate that these efficiencies will temper power and data center demand. In reality, these efficiencies are unlikely to offset the rising compute and energy requirements driven by model complexity, scale and increasing AI adoption.
According to Stanford University's 2025 AI Index Report, leading AI companies have recently shifted their focus from solely scaling models during pre-training to now prioritizing data efficiency during training. This enables smaller models to achieve high performance with less data and lower training costs by training them on smaller high-quality datasets. The chart below compares the smallest models, measured by the number of parameters, capable of scoring above 60% on the Massive Multitask Language Understanding (MMLU) benchmark.
The trend toward compact, high-performing models offers significant benefits, including faster and cheaper inference. This trend reduces the computational and financial barriers for users and organizations to run leading models. Combined with growing datasets, these efficiency gains are expected to enable the development of even more powerful and smaller models in the future, further democratizing AI development and increasing deployment across diverse use cases.
As AI models become more efficient and compute resources grow more accessible, inference costs will continue to decline. This declining cost for compute will accelerate AI adoption, fostering new innovative use cases - and increasing demand for compute. The surge in compute demand following efficiency breakthroughs like DeepSeek exemplifies this trend. According to McKinsey, preliminary analysis of the compute power demand curve agrees that "efficiency gains may not substantially impact overall compute power demand over the long term."
The chart below illustrates inference price trends of some AI models across select benchmarks, showcasing improvements in efficiency and cost reductions that we believe are healthy for the industry and will drive AI adoption at an accelerated pace.
Efficiency improvements that reduce compute costs and enable novel applications underpin the Jevons paradox, where these gains drive increased AI adoption and resource consumption. Research from Artificial Analysis shows that, although efficiency gains have lowered compute requirements for specific tasks, compute demand has outpaced these savings. Enhanced algorithms and training data enable smaller models to achieve greater intelligence, reducing compute needs by approximately one-tenth. Additionally, software and hardware efficiency improvements have cut compute requirements and costs by one-third each. However, scaling laws have driven larger models to require up to five times more compute per query. Innovations, such as reasoning models, are estimated to consume on average 10 times more tokens than single queries (up to 20 times in some cases), and the AI agent revolution is projected to generate 20 times more requests per use. This trend mirrors the Jevons paradox, where decreasing costs lead to significant increases in compute demand and, consequently, power consumption.
The demand for compute in AI workloads continues to surge. Breakthroughs such as DeepSeek and inference-time scaling are accelerating advances in model efficiency, complexity, and usage, driving up compute needs at an exponential pace. This rising demand is outstripping efficiency gains, fueling greater requirements for larger and more power-dense data centers.
Helios is uniquely equipped to address the escalating demand for AI compute infrastructure. The Helios campus has the rare characteristics - reliable power at scale, high bandwidth and low latency fiber access, vast land and water resources, a skilled labor force, and proximity to leading academic institutions - that enable it to support substantial data center capacity growth.
Massive Power Capacity: Helios boasts 800 MW of approved power and an incremental 2.7 GW under study, totaling 3.5 GW of capacity at a single campus that rivals entire major data center markets like Northern Virginia or Atlanta. This extensive energy supply powers large-scale AI data centers and ensures scalability for future AI demands. The ability to train models at a single site that offers this much power can lead to significant efficiency gains and reduce the cost to train, thereby enhancing the economics of model development and monetization for AI companies. Moreover, concentrating massive power capacity at scale allows tenants to optimize infrastructure utilization, streamline operations, and ultimately lower the total cost of ownership across their deployments.
High-Speed Connectivity: With dedicated fiber delivering 10-15 milliseconds (ms) of latency between Helios and Dallas, supported by two diverse paths designed for N+1 redundancy, Helios provides reliable, high-bandwidth performance. This low-latency design enables AI training and inference workloads, ensuring scalability and long-term resilience.
Expansive Land: With control of over 1,500 acres of flat, contiguous land, Helios is poised to grow to be one of the world's largest data center campuses.
Sustainable Water Supply: Located above one of the country's largest aquifers, Helios secures reliable water access to meet the cooling demands of ancillary data center equipment.
Skilled Workforce and Proximity to Academic Institutions: Helios benefits from proximity to a robust labor market in Lubbock, Texas, ensuring access to a skilled workforce capable of supporting advanced data center operations.
The relentless growth in demand for AI compute shows no signs of slowing down. This sustained demand underscores the need for robust, scalable infrastructure to support increasingly power-hungry AI workloads. Helios is uniquely positioned to capitalize on this trend, with cutting-edge capabilities designed to meet the rising requirements for both training and inference. In addition, its extensive access to energy enables Helios to scale out efficiently. Helios stands ready to lead the charge in powering the next generation of innovation.
Legal Disclosure:
This document, and the information contained herein, has been provided to you by Galaxy Digital Inc. and its affiliates ("Galaxy Digital") solely for informational purposes. This document may not be reproduced or redistributed in whole or in part, in any format, without the express written approval of Galaxy Digital. Neither the information, nor any opinion contained in this document, constitutes an offer to buy or sell, or a solicitation of an offer to buy or sell, any advisory services, securities, futures, options or other financial instruments or to participate in any advisory services or trading strategy. Nothing contained in this document constitutes investment, legal or tax advice or is an endorsement of any of the stablecoins mentioned herein. You should make your own investigations and evaluations of the information herein. Any decisions based on information contained in this document are the sole responsibility of the reader. Certain statements in this document reflect Galaxy Digital's views, estimates, opinions or predictions (which may be based on proprietary models and assumptions, including, in particular, Galaxy Digital's views on the current and future market for certain digital assets), and there is no guarantee that these views, estimates, opinions or predictions are currently accurate or that they will be ultimately realized. To the extent these assumptions or models are not correct or circumstances change, the actual performance may vary substantially from, and be less than, the estimates included herein. None of Galaxy Digital nor any of its affiliates, shareholders, partners, members, directors, officers, management, employees or representatives makes any representation or warranty, express or implied, as to the accuracy or completeness of any of the information or any other information (whether communicated in written or oral form) transmitted or made available to you. Each of the aforementioned parties expressly disclaims any and all liability relating to or resulting from the use of this information. Certain information contained herein (including financial information) has been obtained from published and non-published sources. Such information has not been independently verified by Galaxy Digital and, Galaxy Digital, does not assume responsibility for the accuracy of such information. Affiliates of Galaxy Digital may have owned, hedged and sold or may own, hedge and sell investments in some of the digital assets, protocols, equities, or other financial instruments discussed in this document. Affiliates of Galaxy Digital may also lend to some of the protocols discussed in this document, the underlying collateral of which could be the native token subject to liquidation in the event of a margin call or closeout. The economic result of closing out the protocol loan could directly conflict with other Galaxy affiliates that hold investments in, and support, such token. Except where otherwise indicated, the information in this document is based on matters as they exist as of the date of preparation and not as of any future date, and will not be updated or otherwise revised to reflect information that subsequently becomes available, or circumstances existing or changes occurring after the date hereof. This document provides links to other Websites that we think might be of interest to you. Please note that when you click on one of these links, you may be moving to a provider's website that is not associated with Galaxy Digital. These linked sites and their providers are not controlled by us, and we are not responsible for the contents or the proper operation of any linked site. The inclusion of any link does not imply our endorsement or our adoption of the statements therein. We encourage you to read the terms of use and privacy statements of these linked sites as their policies may differ from ours. The foregoing does not constitute a "research report" as defined by FINRA Rule 2241 or a "debt research report" as defined by FINRA Rule 2242 and was not prepared by Galaxy Digital Partners LLC. Similarly, the foregoing does not constitute a "research report" as defined by CFTC Regulation 23.605(a)(9) and was not prepared by Galaxy Derivatives LLC. For all inquiries, please email [email protected]. ©Copyright Galaxy Digital Inc. 2025. All rights reserved.