Thursday, June 12, 2025

Waymo Features a Rider Cost "Premium" Compared to Uber or Lyft, By Design

It appears there is an early adopter pricing premium being paid for Wayno rides, compared to either Uber or Lyft, according to a study by Obi. 


To be sure, that is by design: Waymo entered the market with a premium pricing position. So going driverless doesn’t mean a cheaper ride; instead it is deliberately priced at a premium to either Uber or Lyft. 

source: Obi

Wednesday, June 11, 2025

Why AI Era of Computing is Different

If we can say computing has moved through distinct eras, each with distinct properties, it is not unreasonable to predict that artificial intelligence represents the next era. And though earlier generations are normally defined by hardware, that is less true of more-recent eras, where virtualization is prevalent and the focus is more on applications than hardware. 


But AI might shift matters further. 


Era

Key Feature

Key Technologies

Mainframe Era (1950s–1970s)

Centralized computing

IBM mainframes

Personal Computing Era (1980s–1990s)

Decentralization, personal access

PCs, MS-DOS, Windows

Internet Era (1990s–2000s)

Connectivity, information access

Web browsers, search engines

Mobile & Cloud Era (2000s–2020s)

Always-on, distributed services

Smartphones, AWS, Google Cloud


The AI era should feature software “learning” more than “programming.” Where traditional software follows explicit rules, AI models learn from data, discovering patterns without being explicitly programmed.


AI systems can generalize from experience and sometimes operate autonomously, as in the case of self-driving cars, recommendation systems or robotic process automation.


Voice assistants, chatbots, and multimodal systems mark a transition to more human-centric interfaces, moving beyond keyboards and GUIs.


AI can be considered a distinct era of computing, not because it introduces new tools, but because it changes the nature of computing itself, from explicitly coded systems to systems that evolve and learn.


Tuesday, June 10, 2025

Why Language Models Tackling Very-Complex Problems Often Provide Incorrect Answers

It perhaps already is clear that large language models using “reasoning” (chain of thought, for example) can provide much-better accuracy than non-reasoning models. 


Reasoning models consistently perform well on simple queries. Benchmarks such as MMLU (Massive Multitask Language Understanding) and ARC (Abstraction and Reasoning Corpus) show that models reach near-human or superhuman accuracy on tasks that do not require multi-step or abstract reasoning.


But reasoning models frequently exhibit "reasoning is not correctness" gaps as query complexity grows. Performance degrades sharply as step counts increase, as is required for complex tasks. Token usage grows three to five times for complex queries while accuracy drops 30 percent to 40 percent compared to simple tasks.


Reasoning models using techniques like Chain-of-Thought (CoT) can break down moderately complex queries into steps, improving accuracy and interpretability. However, this comes with increased latency and sometimes only modest gains in factuality or retrieval quality, especially when the domain requires specialized tool usage or external knowledge.


The biggest issues come with the most-complex tasks. Research shows that even state-of-the-art models experience a "reasoning does not equal experience" fallacy: while they can articulate step-by-step logic, they may lack the knowledge or procedural experience needed for domain-specific reasoning, for example. 


That happens because the reasoning models use logically coherent steps that contain critical factual errors. In scientific problem-solving, 40 percent of model-generated solutions pass syntactic checks but fail empirical validation, for example. 


Such issues are likely to be most concerning for some use cases, rather than others. For example, advanced mathematics problems such as proving novel theorems or solving high-dimensional optimization problems often require formal reasoning beyond pattern matching


Problems involving multiple interacting agents (economic simulations, game theory with many players) can overwhelm LLMs due to the exponential growth of possible outcomes.


In a complex negotiation scenario, an LLM might fail to account for second-order effects of one agent’s actions on others. 


Also, problems spanning multiple domains (designing a sustainable energy grid involving engineering, economics, and policy) require integrating diverse knowledge LLMs were not trained on. 


Of course, one might also counter that humans, working without LLMs, might very well also make mistakes when assessing complex problems, and also produce logically reasoned but still "incorrect" conclusions! 


But there are probably many complex queries that still will benefit, as most queries will not test the limits of advanced theorems, economic simulations, game theory, multiple domains and unexpected human behaviors. 


So for many use cases, even complexity might not be a practical issue for a reasoning LLM, even if they demonstrably become less proficient as problem complexity rises. And, of course, researchers are working on ways to ameliorate the issues. 


AI Text Only Suffers from Emotional Flatness In Some Use Cases

As with all generalizations, the claim that writing produced by artificial intelligence or generative AI suffers from a lack of emotion requires some elaboration. Not all writing tasks require or even allow much “emotional expression.”


Academic or essay writing; advertising and marketing content; history or instructional content might do just fine with a straightforward style. 


On the other hand, most of us might wonder how well present and future models will be able to handle fiction, where nuance, emotional depth, and subtlety or a unique voice do matter. 


AI might also not be so great for memoirs, reflective essays, or opinion pieces.


The reasons for the difference are pretty simple. Genres such as academic writing, advertising, and instructional content follow established structures and therefore are easier for AI to mimic.


Fiction and personal narratives require a level of creativity, empathy, and emotional understanding that AI systems currently struggle to replicate. AI can mimic certain tones and styles, but it often lacks the unique voice and perspective that human writers bring to their work.


The point is that AI content, which already is prevalent, will seem more appropriate in some genres, compared to others. One size, as they say, does not fit all. And as useful as AI might be for many humans, in many situations, writers are not going to stop writing because they could use AI for that purpose. 


No, writers write because they enjoy the craft of writing, just as musicians play music or artists paint. AI will not deter any of these creators from doing what they enjoy. My brother wouldn't get any enjoyment out of having AI paint a picture. My sister wouldn't prefer that an AI create and play music. I wouldn't be interested in using AI to write on my behalf. I write because I enjoy the process.


Still, to the extent that AI is a tool to automate or speed writing tasks for many, when it is simply a practical task, the inability to fully mimic human nuance will not be an issue. We don't expect nuance in our emails, ads, marketing copy, technical training manuals or instructional material, really. We never expect it for legal, academic or technical writing, either.


"Lack of emotion" is an issue mostly for creative or fiction writing or biographies; film and TV scripts and often musical lyrics.


Monday, June 9, 2025

How to Avoid or Reduce the Danger of Model Collapse

Recursive training on synthetic data, often referred to as "model collapse," is a significant challenge for artificial intelligence models. It is a point Apple researchers made recently. A paper entitled The Illusion of Thinking suggests reasoning language models underperform standard models for simple problems; show some advantage for medium-complexity tasks but collapse under the weight of high-complexity tasks. 


“We acknowledge that our work has limitations,” the authors note. “While our puzzle environments enable controlled experimentation with fine-grained control over problem complexity, they represent a narrow slice of reasoning tasks and may not capture the diversity of real-world or knowledge-intensive reasoning problems.” 


Some might note that the problem of complex problem processing is not new. In machine learning, recursive training generally refers to a process where a model is trained, and then its own outputs (or outputs from a previous version of itself) are used as part of the training data for subsequent iterations of the model. 


This creates a feedback loop where the model is essentially "learning from itself" or from other models that have similarly learned from generated data.


This leads to a degenerative process where errors compound, and the model's outputs become increasingly narrow, repetitive, and nonsensical.


As always, developers and architects have methods for reducing such distortions. 


The most crucial strategy is to retain a significant portion of original, human-generated, "real" data. By some estimates, even a small percentage of real data (10 percent) can significantly slow down or prevent model collapse.


Continuously introducing new, diverse real data into the training pipeline can counteract the degradation caused by synthetic data. 


Architects also can implement robust processes to verify the quality and accuracy of synthetic data before it's used for training.


Also, architects can focus on creating synthetic data to address specific model "blind spots" or underrepresented scenarios in the real data. 


Architects can also try to ensure that the synthetic data generated is diverse and representative of the desired data distribution, rather than simply mimicking the most common patterns. 


Human feedback also is helpful, and could involve humans evaluating the quality of generated data and providing guidance for the next generation.


Training can use regularization methods (L1/L2 regularization, dropout) during training to prevent overfitting and encourage the model to learn more robust representations. Reinforcement Learning with Human Feedback (RLHF) can be used to align the model's outputs with human preferences, effectively guiding the generation process towards more desirable and accurate results, even when using synthetic data.


The core principle to combat model collapse is to ensure that the model always has access to a reliable source of "ground truth" information, whether through direct inclusion of real data, or through careful curation and validation of synthetic data.


Will Content Companies Fair Better than Telcos with Content Business Separations?

With Warner Brothers Discovery slitting into two companies, one containing the linear video networks and the other the studies and streaming business, some of us might recall similar splits in other industries intended to separate low-growth, cash-flow-generating businesses from higher-growth activities. 


It might seem, in retrospect, an odd thing that telecom executives at AT&T and Rochester Telephone Co., for example, once considered “long distance calling” the growth business, where the local telecom operations were seen as the low-growth entities. 


As it turned out, long distance calling was not much of a business at all, once internet calling developed. Competition was an issue, of course, but the real profit killer was free and nearly-free IP-based calling, plus the rise of mobile calling that simply included domestic long distance calling as a feature. 


Period

Market Structure

Estimated Profit Margin

1985-1990

Post-AT&T Breakup Transition

60-80%

1990-1995

Early Competition

45-60%

1995-2000

Intense Competition

25-40%

2000-2005

Market Maturation

15-30%

2005-2010

Mobile Disruption

5-20%

2010-2015

VoIP/Internet Disruption

0-10%

2015-2020

Legacy Service

-5-5%

2020-2025

Vestigial Market

-10-0%


The point is that, as logical as it seems, such asset separations often do not work out as expected. 



The fundamental problem, some will argue, is that linear TV networks are experiencing accelerating decline rather than stable cash flow. 


source: PwC 


Year

Share of TV Households Without A Traditional TV Connection

2026*

75%

2023

60%

2022

53%

2021

47%

2020

41.70%

2019

36.10%

2018

30.60%

2017

26.20%

2016

22.60%

2015

20%

2014

18.80%


The theory behind separating growth and harvest assets assumes the declining business can generate predictable cash flows while the growth business receives focused investment. But some might note that linear TV's decline appears too steep for reliable cash generation.


The moves might still be “rational” from a management perspective, as it offloads assets in terminal decline (allowing a “harvesting” strategy) while unburdening the potential growth assets (reducing debt, boosting revenue growth rates, reducing overhead). 


Sure, people are going to propose all sorts of new business models or products that might be created. That sort of advice has been common since the 1990s for telecom service providers. And, fortunately, mobility service rose to become the driver of revenue and profits, with the rise of internet access helping replace lost voice revenues as well. 


It might be harder to envision what changes of product or revenue model could develop for linear video.


Sunday, June 8, 2025

Why Apple Might Not Need to "Lead" AI

As Apple gears up for the typically-important Worldwide Developers Conference, many seem uneasy about Apple’s ability to provide evidence that its artificial intelligence strategy (“Apple Intelligence” and Siri) is working. Some worry Apple started too late or is in some ways hobbled because of its emphasis on data privacy, which limits the degree to which Apple software can take advantage of cloud-based processing. 


Those concerns might ultimately be a bit harsh.


Apple, despite the growth of its services revenue, remains a hardware company driven by sales of general-purpose devices such as smartphones. And there is an argument to be made that AI’s value is highest for special-purpose embodiments, less for general-purpose devices. 


In other words, AI’s value is crucial for automated vehicles. We might argue that AI-based health monitors likewise are instances where the AI has moderately-high value. 


For general-purpose smartphones, AI obviously aids photography operations, voice interactions or personalization. But we might expect too much of AI as a value driver beyond that. The specific value of AI for personal computers is arguably even lower.


Sure, voice interfaces are helpful, but most of the location-based personalization smartphones enable is not so pronounced with PCs. One might even argue AI provides lowish value for PCs. 


The point is that Apple’s AI strategy and imperatives are different from those of Alphabet, Microsoft or Meta. There is an argument to be made that Apple primarily has to employ AI to improve the value of phone apps (camera, for example) or user interface. AI is useful, but not existential.


For Microsoft, AI enables many of its core businesses, from gaming to enterprise productivity apps. AI might not emperil its core revenue drivers, so long as it can keep up. For Meta, whose revenue is built on content and advertising, AI might be a net plus.


Amazon might benefit directly mostly through AI-powered logistics efficiencies and recommendations and personalization for its customers.


Alphabet, on the other hand, faces the possible cannibalization of its search business model by AI alternatives. So for Alphabet, AI leadership might be an existential challenge.


So there is an argument to be made that Apple does not actually have to "lead" in broader AI. In fact, that has tended to be Apple's approach to innovation in the past, in any case. It rarely is "first" to introduce a new type of product or category. It tends to "do it better." 

And some of us recall that Apple has faced many periods when it seemed threatened. Eventually, Apple has been able to surmount such challenges. 


Period

Challenges

Rebound Strategy

Outcome

Mid-1980s to 1997

- Declining Mac sales

- Fierce competition from Windows PCs

- Poor leadership after Steve Jobs was ousted in 1985

- Mounting losses and product confusion

- Steve Jobs returned in 1997 through the NeXT acquisition

- Simplified product line

- Microsoft invested $150M in Apple (1997)

- Launched iMac in 1998, designed by Jony Ive

- Apple returned to profitability

- iMac became a major hit

- Re-established design and innovation culture

Early 2000s (2001–2003)

- Skepticism over Apple's entry into consumer electronics

- Slow Mac sales

- Tech bubble aftermath weakened investor confidence

- Launched iPod in 2001

- iTunes Store in 2003 revolutionized digital music

- Strengthened brand loyalty

- iPod became a cultural phenomenon

- Boosted revenue and brand visibility

- Set stage for future devices

2007–2009 (Post-iPhone Launch)

- iPhone faced strong criticism for lacking features (e.g., no 3G, no physical keyboard)

- Doubts about Apple entering the mobile phone market

- 2008 financial crisis hurt tech stocks

- Rapid iteration: iPhone 3G (2008), App Store launch

- Aggressive global carrier partnerships

- Focus on software ecosystem

- iPhone became Apple’s flagship product

- App Store created a new app economy

- Massive revenue growth

2011–2013 (Post-Steve Jobs Era)

- Concerns over Apple’s innovation capacity after Jobs’ death in 2011

- Critics claimed Apple was no longer a “visionary” company

- Increasing Android competition

- Strong product roadmap under Tim Cook

- Continued success with iPhone, iPad, and Mac

- Services revenue growth began (iCloud, App Store, etc.)

- Stock rebounded and reached new highs

- Apple maintained leadership in premium devices

- Cemented Cook’s leadership credibility

2015–2016 (iPhone Saturation Fears)

- Slowing iPhone growth

- China market concerns

- Critics questioned reliance on a single product line

- Diversification: Apple Watch, AirPods

- Expansion of services (Apple Music, iCloud, App Store)

- Focus on ecosystem lock-in

- Apple became world's most valuable company again

- Services and wearables became major revenue contributors

2020 (COVID-19 Pandemic)

- Factory closures and supply chain disruptions

- Retail stores shut down

- Global economic uncertainty

- Rapid pivot to remote work culture

- Launched Apple Silicon (M1 chip) in 2020

- Robust online sales strategy

- Record-breaking quarters post-pandemic

- M1 chip received critical acclaim

- Apple solidified vertical integration strategy 

Waymo Features a Rider Cost "Premium" Compared to Uber or Lyft, By Design

It appears there is an early adopter pricing premium being paid for Wayno rides, compared to either Uber or Lyft, according to a study by Ob...