Cloud Pricing No Longer Works in an AI World

Cloud Pricing No Longer Works in an AI World

Back in my younger years there was a hilarious Saturday Night Live skit about a bank that only made change (You can see it here). It ends with the customers asking – “How do you make money at this?” The answer from the bank was absurd and an LOL moment for this 2nd year MBA student – “It simple. Volume.”

We are entering the 3rd major shift in software pricing in my retail career. When I started over 30 years ago everything was on-prem. Payments up front for everything, often a client license and server license. It was about 10-12 years ago that things began to change to a SaaS model or a pay-as-you go model and hybrid costs for client and usage. This was a huge disruption for existing software providers as they had the classic business “innovators dilemma,” do we destroy our business profitability in the short term for a longer-term benefit. Now it is time for the cloud innovators to be disrupted because the pay as you go no longer works. How quickly will companies adapt?

The emergence and adoption of artificial intelligence (AI) technologies pose a new challenge for cloud pricing models. AI is not just another software application. It is a resource-intensive process that requires massive amounts of computing power, data, and storage. AI workloads can vary significantly in their complexity and duration, making it difficult to predict and budget for their resource consumption. Moreover, AI queries can require up to 10 times more processing power than traditional cloud workloads, resulting in higher and more volatile costs for cloud customers. Thus, all users are not the same when it comes to computing.

Who will make the big money in AI by the time it is over

The Rise of AI and Its Resource Demands

How can cloud providers and SaaS companies adapt their pricing models to the AI era? How can they offer their customers a fair and flexible pricing strategy that aligns with the resource demands of AI? AI workloads can vary significantly in their complexity and duration, depending on the type, size, and quality of the data, the algorithm, the model, and the task.

To illustrate the resource demands of AI, let us look at some examples of AI use cases and their resource requirements. According to a report by Deloitte, a typical natural language processing task, such as sentiment analysis, can require up to 100 gigabytes (GB) of data and 10 gigaflops (GFLOPS – Billions of Floating-Point Operations per Second) of processing power per hour. A typical computer vision task, such as face recognition, can require up to 10 GB of data and 100 GFLOPS of processing power per hour. A typical machine learning task, such as image classification, can require up to 1 terabyte (TB) of data and 1,000 GFLOPS of processing power per hour. These numbers are orders of magnitude higher than the resource requirements of traditional cloud workloads, such as web hosting, email, or database, which can require up to 1 GB of data and 1 GFLOPS of processing power per hour.

These examples show that AI workloads can require up to 10 times more processing power than traditional cloud workloads, and that the resource requirements of AI can vary significantly depending on the use case and the task. This poses a challenge for the traditional cloud pricing models, which are based on fixed and predictable resource consumption. That is just the processing power required, you must also consider the electrical power consumption and data storage as well.

Limitations of Traditional Cloud Pricing Models in the Age of AI

The traditional cloud pricing models are based on the assumption that the resource consumption of cloud customers is fixed and predictable. They offer customers two main options: pay-as-you-go or reserved instances. Pay-as-you-go pricing allows customers to pay only for the resources they use, such as CPU, memory, disk, or network, on an hourly or per-second basis. Reserved instances pricing allows customers to pay upfront for a fixed amount of resources for a fixed period of time, such as one or three years, and receive a discount on the hourly rate.

However, these pricing models have some limitations when it comes to AI workloads. First, they do not account for the variability and unpredictability of AI resource consumption. AI workloads can fluctuate significantly in their resource requirements, depending on the type, size, and quality of the data, the algorithm, the model, and the task. This makes it difficult for customers to accurately predict and budget for their AI resource consumption and exposes them to the risk of cost overruns and unpredictable expenses. For example, a customer who uses pay-as-you-go pricing may end up paying more than expected if their AI workload consumes more resources than anticipated, or if the cloud provider increases the prices of the resources. A customer who uses reserved instances pricing may end up paying more than necessary if their AI workload consumes less resources than reserved, or if the cloud provider lowers the prices of the resources.

Second, they do not offer a granular and flexible pricing approach tailored to AI workloads. AI workloads can have different resource requirements depending on the use case and the task. For example, training an AI model may require more CPU and memory than inferencing an AI model. However, the traditional cloud pricing models do not allow customers to pay for the specific resources they need for their AI workloads and force them to pay for the resources they do not need. For example, a customer who uses pay-as-you-go pricing may end up paying for the CPU and memory they do not use, or for the disk and network they do not need. A customer who uses reserved instances pricing may end up paying for the resources they do not use for the entire duration of the reservation, or for the resources they do not need for the entire duration of the reservation.

What are Retailers Willing to Pay for and when will they Adopt AI options in solutions?

Adding to the complexity is retailers will likely not want to use the AI features for all users and at the same time? The typical cloud solution treats all users as the same. There are users at the store levels that need one level of flexibility, others that need AI all the time, and others in the middle. If vendors try to continue with a one-size fits all approach to licensing, the end result could be devastating. It could be a death knell to their profitability on one side or a risk to the top line revenue on the other as customers push back.

Rethinking Licensing Models for AI

How can cloud providers and SaaS companies rethink their pricing models for AI? How can they offer their customers a pricing strategy that aligns with the resource demands of AI, and that offers them granular and flexible pricing options based on their actual resource usage? We propose a consumption-based pricing model for AI workloads, which is based on the following principles:

  • Granular pricing based on actual resource usage. Customers pay only for the resources they use, such as CPU, memory, disk, or network, on a per-second or per-query basis. Customers can also choose the type and quality of the resources they need for their AI workloads, such as standard, premium, or custom.

  • Pricing tiers or packages tailored to different AI use cases and resource requirements. Customers can choose from different pricing tiers or packages that offer different levels of resources, performance, and features for their AI workloads, such as basic, standard, advanced, or enterprise. Customers can also customize their pricing tiers or packages according to their specific needs and preferences.

  • Transparency and predictability in pricing models. Customers have full visibility and control over their resource consumption and costs and can monitor and manage their AI workloads and budgets in real time. Customers can also estimate their costs and compare different pricing options before committing to a pricing plan. By adopting a consumption-based pricing model for AI workloads, cloud providers and SaaS companies can offer their customers a number of benefits, such as:

  • Cost savings and efficiency. Customers can save money and optimize their resource utilization by paying only for the resources they use, and by choosing the type and quality of the resources they need for their AI workloads.

  • Flexibility and scalability. Customers can adjust their resource consumption and costs according to their changing needs and demands, and scale their AI workloads up or down without any penalties or lock-ins.

  • Customer satisfaction and loyalty. Customers can enjoy a fair and transparent pricing strategy that aligns with their resource demands and usage, and that offers them granular and flexible pricing options tailored to their AI use cases and requirements.

No doubt this will be incredibly challenging and complex at the same time. Vendors may have to use AI for that! But the status quo is simply not going to work. Those that adopt may take a short term hit to profitability but will be the big winners in AI in the future.

What are your thoughts?

IHL Services Jerry Sheldon Beth Buzek Gabriel Buzek Lee Holman 🏷 Ricardo Belmar Vicki Cantrell Cathy Hotka Andrew Laudato Georgina Nelson Justin Honaman David Dorf ShiSh Shridhar

I can stab at answer to this quandary. The cost/benefit equation will stabilize in favor of gradual cost reductions. This situation reminds me patently of circa 1980, working on new POS systems search for the now-defunct Waldenbooks. In those early days of transition away from ECRs, we wanted very much to be able to store an expanded PLU file in each register. There were not many RDMSs around and Oracle was one choice. However, nobody had faced the idea of an RDBMS inside a POS terminal before, and we would need approximately 1,000 licenses. Oracle was licensing their RDBMS for industrial use at approximately $50,000 per instance. We were clearly not going to spend that kind of money and politely told Oracle that they were nuts if they thought that pricing for scale (we were not using hat expression then) was going to fly. Months later they lowered the price to $500 per instance for that purpose only. Similar things will happen with the cost of adopting AI. Many businesses will want to employ it but will resist the currently high cost of ownership, and costs will come down to meet reality.

Wow, this brings back memories of the 1970s and mainframe processing issues at the computer science lab at college. The choke point was CPU time. Each computer class had a control number and you would put that number on the fourth card of your IBM punch card deck. The deck would be run through a reader and your job would be in the queue. If it was a quiet day and no one was at the lab, your job would get processed right away and your printout would appear on the shelf in a few minutes. At the end of the term, the facility would be very busy and your job might not get back for 6 or 8 hours. If your class ran out of CPU time the job would not run, it would come back with a message that the account was empty and you had to let the professor know and they would get more time added. Or, you might just copy the fourth card off a deck that had recently run and was waiting on the shelf, belonging to some other computer class, and resubmit your job, thus billing it to some other class. I think time of day processing could be a way to price CPU time, along with other services. Does every query need to be processed immediately? Or can a few queries be submitted at 4:30pm for turnaround by 9am the next day? Answers will be found.

Scott Phillips

Chicago-based PR firm founder, communications strategist serving leading B2B corporations and other organizations

1mo

Spot-on analysis, Greg. Thanks for this.

This is a great article as it exposes that which most people don't understand when all they hear is AI does this or that. The real cost of AI is expensive. This video explains the AI process in layman's language. "Artificial Intelligence in Retail" https://www.youtube.com/watch?v=_3bDFpEKrK0&t=20s Thanks for more great stuff Mr. Buzek.

To view or add a comment, sign in

Insights from the community

Others also viewed

Explore topics