Tuesday, October 31, 2023

How Will Most Firms Recover AI Usage Costs?

Since most users of cloud-based software-as-a-service services already are accustomed to pricing based substantially on usage, generative AI might represent few new pricing issues. If their customers use more, they will pay more.


For cloud computing suppliers, there are at the moment perhaps fewer issues about how to charge for usage than about how to create the high-performance compute fabric AI in general and large language models in particular require. 


Entirely different issues confront most other firms that must figure out how to price AI capabilities incorporated into their products. For most firms, AI is a new cost that must be recovered somehow in retail prices charged to customers.


Uncertainty about levels of usage is one variable. But there is no uncertainty about product costs when AI features are based on usage of "cloud computing as a service." Inference operations are going to require cloud computing usage, each time an inference operation is invoked.


How to recover the costs of paying for cloud compute therefore is a new question to be answered.


For most firms that will want to use large language models (generative AI), the big issue is how to recover the cost of LLM features used by their customers. So far, the most-common models are:

  • AI is a feature of a higher-priced version of the existing product (higher-cost plan versus standard)

  • AI is a value-added feature with an extra flat-fee cost

  • AI is a feature of an existing product for which there is no direct incremental charge to the user (such as a search or social media or messaging user), but might eventually represent a higher cost to customers (advertisers or marketers buying advertising, for example)

  • AI is a no-charge feature of an existing product, but with usage limits (freemium)

  • AI is a new product with charges that are largely usage-based (GenAI “as a service” offered by infrastructure-as-a-service providers). 


And some software firms might use a few of those models. For example, Microsoft charges for its AI-assisted copilots, including those in Office and GitHub, with prices ranging from $5 to $40 per user per month. 


But some copilots are included with certain enterprise subscriptions, while a number of Microsoft's consumer AI services remain free for now.


Other software product suppliers also must grapple with how to recover costs of supporting AI features used by their customers.


Box includes AI features for business customers subscribed to its Enterprise-Plus tier and above. Each user will have access to 20 queries per month, with 2,000 additional queries available on a company level. Additional usage will require further payment.


Adobe is including  "generative credits" with its various Creative Cloud, Express and Firefly services.  Starting November 2023, Adobe will offer additional credits using a subscription plan, with plans starting at $4.99 per month for 100 credits. 


“Usage” seems to be the area where there is most danger for retailers, who must make key assumptions about the value of AI when embedded into core products, as well as the cost recovery mechanism when suppliers are not yet sure about how much their customers will use the AI features. 


The key danger will be underestimating usage, unless usage is part of the customer AI pricing formula. 


In a market where retail customers use their own hardware, that would not be an issue. 


But in a market reliant on cloud computing, where retail customers use the supplier’s cloud computing resources, usage really does matter, whenever the supplier is in turn paying a cloud services vendor for compute. 


A few hyperscale cloud computing firms (Microsoft, Google, Facebook) will be somewhat protected, as they can use their own infrastructure. But most enterprises will have to pay retail rates for computing services, so volume does matter. 


Although”compute as a service” suppliers are going to face customer pushback as AI compute loads and charges mount up, at least they tend to be protected as most of their services are usage based. Customers who use more, pay more. 


Businesses that buy “compute as a service” will have to take usage into account. 


Some of those “customer usage and customer pricing” issues might be reminiscent of issues connectivity providers faced in the past as core product “usage-based pricing” shifted. 


Though both flat-fee and usage-based pricing was common in the era where voice was the dominant product, flat fee has been the bigger trend for internet access, interconnection and transport. Within some limits, internet access, for example, tends to be “flat fee” based. 


That poses key issues as the volume of usage climbs, but revenue does not. One can see this in network capital investment, for example, where network architects must assume perpetual 20-percent (or higher) increases in usage every year. 


In some ways, suppliers that embed AI into their products are going to face similar problems. Though cloud computing suppliers will still largely be able to employ usage mechanisms, many retailers of other products are as-yet unclear about how much usage will eventually happen.


That, in turn, means they are as-yet unsure about long-term cost recovery mechanisms and retail AI pricing. 


Flat-fee pricing will be the simplest solution for the moment, and likely the least-objectionable method from a customer standpoint. Whether that continues to work so well in the future is the issue, if AI inference operations grow in volume as much as some might suspect. 


It will be difficult for most firms to sustain low flat-fee rates as volume escalates. The exceptions are those handful of firms that own their cloud compute infrastructure (Microsoft, Google, Facebook, Amazon, for example). 


Of course, some of those sorts of firms will be able to justify “no fee to use” as well, since they have commerce or advertising revenues supporting many of their core products. That is a luxury few firms will experience.


AI usage is going to be a big issue for most firms. So is the issue of how to recover costs related to supplying that usage.

Monday, October 30, 2023

Premature AI Regulation Could Slow its Development

Premature regulation of artificial intelligence--as compelling as it will seem to policymakers--might stifle or slow deployment and benefits, as some have argued has been the case for new technology in many other instances. But the evaluation of such rules often depends on whether it is “consumer advocates” or “industry advocates” who do the viewing. 


Always, “consumer protection” has been viewed as an easy win by lawmakers, who can say they are acting in defense of citizens. But outcomes sometimes are not what lawmakers envisioned. 


The Fairness Doctrine was supposed to support “differing views” on controversial matters. Instead, the rules mostly ensured that broadcasters would avoid such issues. Other rules relating to electronic media likewise have been touted as protections, but also were challenged on grounds of restricting freedom of expression. 


Advocates of network neutrality have argued that the Telecommunications Act of 1996, designed to promote innovation in communications, had unintended consequences for network neutrality, as the Act “might” have allowed internet access providers to block content, favor their own content or extract payments from app providers. 


Opponents of network neutrality say there is scant evidence that such abuses ever happened, even in the absence of rules. Even those who believe ISPs would behave badly if they could might also agree that robust competition seemingly prevents ISPs from acting in such a manner. 


Technology

Regulation

Impact

Motion pictures

The Hays Code (1930-1966)

The Hays Code was a self-censorship code that restricted the content of Hollywood films. The code was designed to uphold traditional moral values, but it also stifled creativity and innovation in the film industry.

Television

The Fairness Doctrine (1949-1987)

The Fairness Doctrine was a policy that required broadcasters to present opposing views on controversial issues of public importance. The doctrine was designed to ensure that the public had access to a variety of viewpoints, but it also stifled free speech and debate.

Video games

The Entertainment Software Rating Board (ESRB) (1994-present)

The ESRB is a self-regulatory body that rates video games based on their content. The ESRB was created in response to public concerns about the violence in video games. While the ESRB has been successful in reducing the amount of violence in video games, it has also been criticized for stifling creativity and innovation in the video game industry.

Internet

The Communications Decency Act (CDA) (1996-1997)

The CDA was a law that attempted to restrict access to indecent and obscene material on the internet. The law was challenged in court and ultimately struck down as unconstitutional. However, the CDA had a chilling effect on free speech online.

Blockchain

The Securities and Exchange Commission (SEC) (2017-present)

The SEC has been criticized for its regulation of the blockchain industry. The SEC has taken a number of actions against blockchain companies, including issuing cease-and-desist orders and filing lawsuits. The SEC's actions have created uncertainty in the blockchain industry and have made it difficult for blockchain companies to raise capital.

AI

The European Union's General Data Protection Regulation (GDPR)

The development of new AI technologies, such as facial recognition and voice assistants

Internet

The Telecommunications Act of 1996

Early internet development, including the growth of social media and streaming services

Radio

The Radio Act of 1927

Early radio broadcasting, including experimental and avant-garde programming


Sunday, October 29, 2023

What's "Average" in Home Broadband?

As with most anything relating to the internet, “median” or “average” figures often do not reveal as much as we believe, since a relatively small percentage of heavy users exists with larger percentages of light users, along with lots of users somewhere in the middle. 


As a  broad generalization, we might characterize user behavior for any particular app, internet usage in general or data consumption as coming in three buckets.


Consumption

Percentage of users

Heavy

20%

Average

60%

Low

20%


According to data from OpenVault and Ookla, which track internet usage data from millions of broadband customers around the world, heavy users in the U.S. market are defined as those who consume more than 1.2 terabytes of data per month.


“Average” users consume between 300 gigabytes and 1.2 TB of data per month, while light users consume less than 300 GB of data per month. 


As an approximation, home broadband consumers also purchase different tiers of service based on their consumption profiles. The price variance generally is highest for the “heavy” users, where the cost of 1-Gbps service commonly ranges up to about $90 a month, with higher speeds of 2-Gbps, for example, up in the $110 to $125 per month range. 


Consumption

Speed tier

Monthly price

Heavy

2 Gbps or faster

$100-$150

Heavy

1 Gbps

$75-$100

Average

500 Mbps

$50-$75

Low

250 Mbps

$25-$50


Current pricing levels cites by big U.S. internet service providers tends to support the notion of “average” levels of demand with average prices for supplied service. 


AT&T Home Broadband ARPU

Q3'22

Q4'22

Q1'23

Q2'23

Q3'23

Y/Y Growth

Fiber

$62.62

$64.82

$65.92

$66.70

$68.21

8.93%

Non-Fiber

$54.80

$55.54

$56.00

$56.71

$60.43

10.27%

Total Broadband ARPU

$58.63

$60.31

$61.31

$62.26

$64.91

10.71%



Quarter

Comcast ARPU

Charter ARPU

Year-over-year growth rate

Q3 2022

$80.67

$66.10

N/A

Q4 2022

$82.34

$67.75

2.5%

Q1 2023

$83.91

$69.40

4.9%

Q2 2023

$85.48

$71.05

7.4%

Q3 2023

$87.05

$72.70

9.9%


Will AI Fuel a Huge "Services into Products" Shift?

As content streaming has disrupted music, is disrupting video and television, so might AI potentially disrupt industry leaders ranging from ...