IT Brief Australia - Technology news for CIOs & IT decision-makers
Story image
Apples for apples: Measuring the cost of public vs. private clouds
Thu, 26th Oct 2023

How can strategic choices be made on where to place workloads without a consistent comparison?

If the analysts are correct, we’re at or nearing an inflection point. 

The point at which more than half of Enterprise IT spend is made in the cloud. It’s no surprise, then, that an increasing focus among IT leaders is on both the absolute cost and the effectiveness of that cloud spend within their organizations. For some, this has been more than a challenge, with one of the very reasons that cloud has become so popular in the first place - the ability for engineers to swipe a credit card and spin up effectively limitless resources, enabling them to commit to elastic levels of spending, often under the radar of any central budgetary control. This has hampered many in their ability to both get a handle on exactly what and where these costs are coming from and also to plan effectively for a more strategic relationship with the hyperscalers. 

Enter FinOps and the Cloud Centres of Excellence

A portmanteau of “Finance” and “DevOps” entered the relatively recent discipline of “FinOps”. FinOps is a collection of best practices and tooling with the ultimate aim of optimizing the business value derived from spending on the cloud. Originally stemming from user group meetings for cloud cost optimization platform Cloudability (now part of IBM, after their recent purchase of Cloudability acquirer Apptio), the FinOps practice is now being moved forward by the independent FinOps Foundation, a project within the Linux Foundation.

While FinOps might provide the cost optimization piece of the puzzle, in parallel, it has given rise to what are being termed Cloud Centres of Excellence (CCoEs), reportedly now operating at either a formal or informal level in the majority of larger organizations and in the process of being rapidly adopted elsewhere. These CCoEs bring a strategic lens to the topic of cloud strategy, considering technical, financial, regulatory and wider business considerations in the longer-term relationship of the organizations with the cloud providers the organization chooses to engage with. 

Comparing “apples to apples” needs FOCUS 

Underpinning all of this at a very practical and tactical level is the ability to ascertain exactly what your organization is actually doing in each cloud provider, how both much it is costing and why the costs are the way they are. At the same time, if the objective is not just to optimize spend and value within a single cloud provider and instead compare the relative benefits of moving to a different cloud provision, then there must be a consistent method for comparison. 

It’s something, according to Forrester’s Tracy Woo, customers have been asking for from the hyperscalers for around a decade, but until very recently, these cries had not been met. 

That’s the objective behind a recent open-source standards project called FOCUS: the FinOps Open Cost and Usage Specification. FOCUS is a Linux Foundation technical project sponsored by the FinOps Foundation, with the first official release slated for the end of 2023.  This project will define a specification that allows cost and usage data to be produced using a common schema and using “FinOps-compatible” terminology. It should then be possible to easily unify relevant billing data across all providers and drive better business decisions based on the fully burdened cost of using the cloud.

But what about private cloud?

Well, for the moment, at least, this seems to be the catch. 

As it appears to me, at least with FinOps as a whole today, the focus of FOCUS seems very much to be on solving the problem within the public cloud providers. It’s not only a multi-cloud world out there, with almost all large organizations working with multiple public cloud vendors today, but more importantly, in the context of this topic, around three-quarters of these organizations also run private clouds, or some “Frankencloud” approximation to one.

The question then is clear. If the objective is to make a true “apples to apples” comparison to decide the optimal location for workloads to be placed, then that analysis is incomplete without considering fully the efficient utilization of assets likely already bought and paid for, consuming space, power and cooling in your own data centre or co-lo. Those assets are there, whether you utilize them at 20% or 80% of their capacity. The incremental cost of running workloads there, over and above the ones you already have to run on-prem., may be practically zero.