Since most users of cloud-based software-as-a-service services already are accustomed to pricing based substantially on usage, generative AI might represent few new pricing issues. If their customers use more, they will pay more.
For cloud computing suppliers, there are at the moment perhaps fewer issues about how to charge for usage than about how to create the high-performance compute fabric AI in general and large language models in particular require.
Entirely different issues confront most other firms that must figure out how to price AI capabilities incorporated into their products. For most firms, AI is a new cost that must be recovered somehow in retail prices charged to customers.
Uncertainty about levels of usage is one variable. But there is no uncertainty about product costs when AI features are based on usage of "cloud computing as a service." Inference operations are going to require cloud computing usage, each time an inference operation is invoked.
How to recover the costs of paying for cloud compute therefore is a new question to be answered.
For most firms that will want to use large language models (generative AI), the big issue is how to recover the cost of LLM features used by their customers. So far, the most-common models are:
AI is a feature of a higher-priced version of the existing product (higher-cost plan versus standard)
AI is a value-added feature with an extra flat-fee cost
AI is a feature of an existing product for which there is no direct incremental charge to the user (such as a search or social media or messaging user), but might eventually represent a higher cost to customers (advertisers or marketers buying advertising, for example)
AI is a no-charge feature of an existing product, but with usage limits (freemium)
AI is a new product with charges that are largely usage-based (GenAI “as a service” offered by infrastructure-as-a-service providers).
And some software firms might use a few of those models. For example, Microsoft charges for its AI-assisted copilots, including those in Office and GitHub, with prices ranging from $5 to $40 per user per month.
But some copilots are included with certain enterprise subscriptions, while a number of Microsoft's consumer AI services remain free for now.
Other software product suppliers also must grapple with how to recover costs of supporting AI features used by their customers.
Box includes AI features for business customers subscribed to its Enterprise-Plus tier and above. Each user will have access to 20 queries per month, with 2,000 additional queries available on a company level. Additional usage will require further payment.
Adobe is including "generative credits" with its various Creative Cloud, Express and Firefly services. Starting November 2023, Adobe will offer additional credits using a subscription plan, with plans starting at $4.99 per month for 100 credits.
“Usage” seems to be the area where there is most danger for retailers, who must make key assumptions about the value of AI when embedded into core products, as well as the cost recovery mechanism when suppliers are not yet sure about how much their customers will use the AI features.
The key danger will be underestimating usage, unless usage is part of the customer AI pricing formula.
In a market where retail customers use their own hardware, that would not be an issue.
But in a market reliant on cloud computing, where retail customers use the supplier’s cloud computing resources, usage really does matter, whenever the supplier is in turn paying a cloud services vendor for compute.
A few hyperscale cloud computing firms (Microsoft, Google, Facebook) will be somewhat protected, as they can use their own infrastructure. But most enterprises will have to pay retail rates for computing services, so volume does matter.
Although”compute as a service” suppliers are going to face customer pushback as AI compute loads and charges mount up, at least they tend to be protected as most of their services are usage based. Customers who use more, pay more.
Businesses that buy “compute as a service” will have to take usage into account.
Some of those “customer usage and customer pricing” issues might be reminiscent of issues connectivity providers faced in the past as core product “usage-based pricing” shifted.
Though both flat-fee and usage-based pricing was common in the era where voice was the dominant product, flat fee has been the bigger trend for internet access, interconnection and transport. Within some limits, internet access, for example, tends to be “flat fee” based.
That poses key issues as the volume of usage climbs, but revenue does not. One can see this in network capital investment, for example, where network architects must assume perpetual 20-percent (or higher) increases in usage every year.
In some ways, suppliers that embed AI into their products are going to face similar problems. Though cloud computing suppliers will still largely be able to employ usage mechanisms, many retailers of other products are as-yet unclear about how much usage will eventually happen.
That, in turn, means they are as-yet unsure about long-term cost recovery mechanisms and retail AI pricing.
Flat-fee pricing will be the simplest solution for the moment, and likely the least-objectionable method from a customer standpoint. Whether that continues to work so well in the future is the issue, if AI inference operations grow in volume as much as some might suspect.
It will be difficult for most firms to sustain low flat-fee rates as volume escalates. The exceptions are those handful of firms that own their cloud compute infrastructure (Microsoft, Google, Facebook, Amazon, for example).
Of course, some of those sorts of firms will be able to justify “no fee to use” as well, since they have commerce or advertising revenues supporting many of their core products. That is a luxury few firms will experience.