Friday, March 21, 2025

Claude Adds Web Search Feature: What it Means

Claude now has added web search for  all paid Claude users in the United States, while support for users on free plans and in more countries is coming soon.


For Claude AI engine’s users, the new feature means up-to-date information beyond its training cutoff is added. So Claude should be more useful for real-time information needs. Users will not see messages that the provided information is only available through the end of 2023, for example, as is common for any engine without real-time search. 


Comparison: Search Engines vs. AI Assistants With and Without Web Search

Capability

Traditional Search Engines

AI Assistants with Web Search

AI Assistants without Web Search

Information Retrieval

Comprehensive access to indexed web content

Access to recent information beyond training data

Limited to information available in training data

Information Recency

Real-time updates and fresh content

Near real-time information (depending on search integration)

Limited by knowledge cutoff date

Result Presentation

List of relevant links requiring user navigation

Synthesized information with cited sources

Synthesized information from training data only

Complex Queries

Keyword-based with some natural language support

Natural language understanding with contextual web results

Natural language understanding limited to training knowledge

Source Transparency

Direct links to original sources

Can cite sources when providing information

Cannot reliably cite specific sources

Factual Accuracy

Varies by source quality; user must evaluate

Generally improved with access to current information

May provide outdated information or hallucinate when uncertain

Personalization

Based on search history and user data

Conversation context + web results tailored to query

Conversation context only

Multi-turn Interaction

Limited (requires new searches)

Strong conversational memory with updated information

Strong conversational memory with static knowledge base

Content Creation

Limited to search result presentation

Creative content informed by latest trends and data

Creative content limited to training data knowledge

Specialized Knowledge

Excellent for finding niche information

Can discover and explain specialized current topics

May struggle with niche or recent specialized topics

Exploration Breadth

Excellent for discovering diverse viewpoints

Can access diverse viewpoints but may synthesize

Limited to viewpoints represented in training data

User Effort Required

Higher (must sift through results)

Lower (provides direct answers with sources)

Lower (provides direct answers)


So Claude becomes more directly competitive with other AI assistants such as ChatGPT and Google's Gemini that already have search capabilities, at least in the paid versions for ChatGPT. It already is t he case that most of the AI assistants, at least in their paid versions, support search, so that feature is becoming table stakes.


Still, some will argue that Claude (as arguably is the case for many other AI assistants) is unlikely to fully substitute for traditional search engines. Search engines arguably remain better for broad exploration and discovery, while AI assistants are better for focused questions.


Search engines provide comprehensive results lists users can browse, while Claude provides synthesized information (even if some AI assistants provide the equivalent of footnotes showing where they got the information. So there are use cases where source attribution is important, and search engines might still be preferable sources.  


Still, adding web search should improve answers about current events with greater accuracy. The likelihood of hallucinations should also be reduced.


Thursday, March 20, 2025

"AI Edge Computing" is Multiple Markets, Not One

AI edge computing refers to the deployment of artificial intelligence algorithms at the "edge" of a network, closer to where data is generated, rather than relying solely on centralized cloud infrastructure. 


But there is a huge difference between “on-device” and “at a remote site” implementations, value chains and markets. 


For example,, on-device edge AI is about smartphones, IoT sensors, wearables or autonomous vehicles. 


Remote edge AI is about data centers, cloud computing, servers and other “enterprise” or “business” computing functions. 


Lumping everything together in one big “AI edge computing” category obscures as much as it illuminates. 


Category

Metric

On-Device Edge AI

Remote Edge AI

Source/Assumption

Market Size (2025)

Financial (USD Billion)

$15 billion

$10 billion

Based on edge AI market growth (e.g., Grand View Research, 21.7% CAGR from $20.78B in 2024)

Market Size (2030)

Financial (USD Billion)

$50 billion

$35 billion

Extrapolated from "device" and "data center" forecasts

Usage (2025)

Devices/Deployments

20 billion devices (smartphones, wearables)

500,000 edge nodes (e.g., servers, gateways)

Statista IoT, IDC edge spending forecasts

Compute Cycles (2025)

Avg. Cycles per Task

10^6 cycles (lightweight models, e.g., NLP)

10^9 cycles (complex models, e.g., video analytics)

Hardware capability estimates

Financial Implication

Revenue Driver

Hardware sales (AI chips, $500 billion smartphone market)

Infrastructure and services , perhaps $450 billion per year

On-device chips, smartphones for "on-device" mkt., connectivity and servers for "remote"

Growth Rate (2025-2030)

CAGR

27%

23%

Higher consumer device adoption vs. enterprise


As you can see, “edge AI” markets are largely contained within other existing device and data center markets. Looking at chip content alone, in either edge device or data center markets is helpful, but doesn’t show the full value chain for either type of product.


Wednesday, March 19, 2025

Some Industries are AI-Resistant

Sometimes conventional wisdom can be quite wrong. Consider the notion that artificial intelligence will disproportionately disrupt jobs in some industries such as agriculture, hospitality and construction. 


At least for the moment, these are some of the areas in which observers expect to see relatively lesser impact from wider use of AI, often in the form of robotic processes. 


And at least for the present, the actual service provided by a human is preferable to that provided by a robot. In many other cases robots cannot cost effectively handle complicated or relatively non-routine use cases that humans can manage easily. 


In fact, jobs such as data entry; transportation; customer service and manufacturing; for example, are considered more likely to be affected by applied AI. 


Other jobs in agriculture; hospitality and construction are viewed as less disrupted by AI, often because the jobs hinge on either human dexterity and customization; fine motor skills, human interaction, empathy and personalized service. 


Job Category

Likely to be Most Affected

Likely to be Least Affected

Rationale

Data Entry & Processing

Data Entry Clerk, Telemarketer, Transcriptionist, Proofreader


AI excels at repetitive tasks, pattern recognition, and data analysis. These jobs are heavily reliant on those skills.

Transportation & Logistics

Truck Driver, Taxi Driver, Delivery Driver, Dispatcher


Self-driving technology is rapidly advancing, potentially automating many transportation roles.

Customer Service

Customer Service Representative, Help Desk Technician, Chatbot Support


AI-powered chatbots and virtual assistants can handle routine customer inquiries, freeing up human agents for more complex issues.

Manufacturing & Production

Assembly Line Worker, Machine Operator, Quality Control Inspector


Automation has been present in manufacturing for a while. AI can further optimize processes, predict maintenance needs, and improve quality control.

Finance

Financial Analyst, Accountant, Tax Preparer, Investment Advisor (entry-level)


AI can analyze vast amounts of financial data, identify trends, and automate routine financial tasks.

Legal

Paralegal, Legal Secretary, Document Reviewer


AI can assist with legal research, document review, and contract analysis.

Creative & Artistic

Graphic Designer (basic), Content Writer (repetitive), Musician (algorithmic generation)


AI is making inroads in generating creative content, though the extent of its long-term impact is still debated.

Skilled Trades


Plumber, Electrician, Carpenter, Welder

These jobs require physical dexterity, problem-solving in unpredictable environments, and often hands-on customization, which are currently challenging for AI.

Healthcare


Surgeon, Nurse, Physical Therapist, Mental Health Professional

While AI can assist in diagnosis, treatment planning, and drug discovery, the human element of care, empathy, and complex decision-making in unpredictable situations is likely to remain crucial.

Management & Leadership


CEO, Manager (complex team dynamics), Entrepreneur

These roles require strategic thinking, interpersonal skills, complex problem-solving, and adaptability, which are currently difficult for AI to replicate fully.

Education


Teacher (personalized instruction, mentorship), Professor

While AI can assist with grading and personalized learning platforms, the human interaction, mentorship, and ability to adapt to individual student needs are likely to remain important.

Science & Research


Scientist (innovative research), Researcher (complex experiments)

AI can accelerate research by analyzing data and identifying patterns, but the creative and critical thinking of human scientists is still essential for breakthroughs.

Tuesday, March 18, 2025

AI Chip Markets and Operations Shifting to "Inference?"


 

The artificial intelligence market changes fast, and not only because new models have been popping up. 

It seems we already are moving towards inference operations as the driver of much of the chip market, for example. 

 As AI adoption scales, cloud and data center operations will prioritize inference-driven AI workloads. 

That will highlight a growing need for specialized hardware optimized for inference tasks, and that arguably is where large end users (Amazon Web Services, Google Cloud, Meta and others) have been working to create homegrown solutions. AWS and Google Cloud, for example, have invested heavily in developing their own AI accelerators, specifically designed for inference tasks. 

The AWS Inferentia is purpose-built for AI inference workloads. Google Cloud Tensor Processing Units are specifically designed for AI workloads, including inference. 

Inference might already represent up to 90 percent of all machine learning costs, for example. 

And lots of capital is being invested in startups aiming to improve processing efficiency.

"Lean Back" and "Lean Forward" Differences Might Always Condition VR or Metaverse Adoption

By now, it is hard to argue against the idea that the commercial adoption of “ metaverse ” and “ virtual reality ” for consumer media was in...