Wednesday, February 19, 2025

Anthropic Plans Reasoning Methods That Vary Computational Intensity

Anthropic (backed by Amazon and Alphabet) is expected to soon introduce a model that uses a slightly different approach to reasoning, allowing designers to tweak the computational effort of the model (essentially, high, medium or low), resulting in differences in how long and how much effort the model puts into reasoning about a particular problem. 


Anthropic seems to be seeking a higher profile as an “enterprise” or “business” model supplier, whose products excel at the sorts of coding larger businesses require. For example, the new model is said to be “better at understanding complex code bases built from thousands of files and produces complete lines of code which work the first time.


That business user focus might also explain why Anthropic is putting effort into features that give developers more control over cost, speed and pricing. 


The model uses more computational resources to calculate answers to hard questions. The AI model can also handle simpler tasks faster without the extra work, by acting like a traditional large language model, or LLM.


The new model might be important for other reasons. One advantage DeepSeek has apparently demonstrated is the way it can reason and learn from other models. 


As was always to be expected, in a fast-moving AI field, important innovations by any single provider are going to be mimicked by other leading contenders as well. 


One might well argue that Anthropic’s new model will provide an example of that, perhaps also illustrating the fact that the DeepSeek approach to reasoning also has been under development or investigation by multiple developers, to some extent. 


No comments:

How Much Will AI Compute Grow to 2030, Compared to Electricity Consumption?

A report by the International Energy Administration estimates data centers accounted for around 1.5 percent of the world’s electricity cons...