Can general-purpose language models run locally on AI PCs? Yes, in “small language model” form. Will more of that be happening in the future? Yes. Is that capability generally useful for most PC users today? No.
But the trajectory is clearly in that direction, essentially shifting more capabilities from “cloud access” to “onboard” processing over time.
The small language model landscape in 2026 has three practical buckets:
ultra-compact models (500M–2B parameters) that run on smartphone processors with 1–4GB RAM
compact models (2B–5B parameters) that handle complex reasoning and coding on consumer hardware
larger efficient models approaching frontier capability.
Most laptop users will not find local processing a huge help for most use cases. It remains unclear how much value local live translation; autocomplete for text; email summarization; note taking or voice assistants provide.
Creative professionals arguably might see the most tangible gains right now:
Adobe Photoshop uses the NPU for Generative Fill, intelligent selection, and automatic retouching
Adobe Premiere Pro's AI features leverage NPUs for scene detection, auto-reframe, and speech-to-text. A 10-minute 4K timeline that previously required 8 minutes for AI analysis now completes in 2 minutes on NPU-equipped systems, while the GPU remains free for color grading. OrdinaryTech
Adobe’s Lightroom Classic uses the NPU for AI-assisted noise reduction in RAW files, and Capture One benefits for automatic cropping and look equalization across large batches of images.
Over time, more-complex tasks could shift on onboard, though. Document creation or some code generation seem likely examples. Gaming and some business productivity use cases also seem likely to benefit.
Over time, more-complex tasks could shift on onboard, though. Document creation or some code generation seem likely examples.
But tasks requiring real-time world knowledge, frontier-scale reasoning or large model access will remain cloud based. Battery life issues might push users to continue using remote solutions, evenif local processing is possible.
The sweet spot for AI PCs over the next few years might be privacy-sensitive, latency-critical or frequently-repeated tasks that have the same sort of economics as any “local hardware versus remote service” tradeoff would feature.