Dedicated GPU infrastructure can beat the public cloud on cost. Companies considering purchasing an AI cluster need to consider utilization as the key variable in their calculations.
Dr. Owen Rogers is Uptime Institute’s Senior Research Director of Cloud Computing. Dr. Rogers has been analyzing the economics of cloud for over a decade as a chartered engineer, product manager and industry analyst. Rogers covers all areas of cloud, including AI, FinOps, sustainability, hybrid infrastructure and quantum computing.
orogers@uptimeinstitute.com
Dedicated GPU infrastructure can beat the public cloud on cost. Companies considering purchasing an AI cluster need to consider utilization as the key variable in their calculations.
Uptime Intelligence looks beyond the more obvious trends of 2025 and examines some of the latest developments and challenges shaping the data center industry.
Cloud providers need to win AI use cases in their early stages of development. If they fail to attract customers, their AI applications may be locked-in to rival platforms and harder to move, which can have serious repercussions.
Visibility into costs remains a top priority for enterprises that are consuming cloud services. Improving the tagging of workloads and resources may help them to spot, and curb, rising costs.
The cost and complexity of deploying large-scale GPU clusters for generative AI training will drive many enterprises to the public cloud. Most enterprises will use pre-trained foundation models, to reduce computational overheads.
While the aim of FinOps is to manage just the cloud costs, technology business management seeks to aggregate all costs of IT, including data centers, servers, software and labor, to identify savings and manage return on investment.
Enterprises have various options on how and where to deploy their AI training and inference workloads. This report explains how these different options balance cost, complexity and customization.
To meet the demand driven by AI workloads, a new breed of cloud provider has emerged, delivering inexpensive GPU infrastructure as a service. Their services are highly demanded today, but longer-term, the market is ripe for consolidation.
While GPUs are the power-hungry devices that enable effective AI training, it is innovations in software that are fueling the recent surge in interest and investment. This report explains how neural networks power generative AI.
Although quantum computing promises a revolution in scientific discovery, its use is still constrained to research and continuing development. However, a new IBM quantum data center in Germany signals a growing interest in its capabilities.
Reserved instances are a pricing model for virtual machines offered by cloud providers. As they offer savings of up to 70% compared with on-demand pricing, organizations should use them liberally, especially in challenging times.
The key benefit of cloud computing lies in its on-demand pricing model. This enables organizations to grow or shrink their applications at will without giving the cloud provider any advance notification. Cloud providers can only offer such…
The Uptime Institute Global Data Center Survey highlights the experiences and strategies of data center owners and operators in areas of resiliency, sustainability, efficiency, staffing, cloud and innovative technologies.
In recent conversations with both regulators and some enterprises, a concept borrowed from the financial sector has been discussed with growing frequency: concentration risk. In finance, the term refers to the level of risk arising from the…
Organizations encounter a bewildering assortment of cloud storage platforms. The difference between the offerings lies in who is responsible for scaling, resiliency and performance: the provider or the customer.