Will edge-based inference kill AI data centers?
If super-efficient language models move to devices, what happens to the trillion-dollar AI data center build out?
I saw a tweet last night which piqued my interest. The tweet claimed that DeepSeek has created a new technique that makes LLMs 60x more efficient than extant ones. In other words, small, powerful models might soon sit on your device: inference moves to the edge.
Of course this raises a somewhat uncomfortable question: if small, device-bound models are sufficiently powerful and capable such that inference moves to devices, what happens to the multi-trillion dollar AI data center build out? What happens to Nvidia? Will Jensen have to sell timeshares in Florida to pay the bills?

You can read DeepSeek’s explanation here. Here’s what they seem to be claiming:
They take long texts, such as paragraphs, multi-page documents, or PDFs, and convert hem into a 2D visual mapping rather than a pure token/text sequence. The model consumes the visual representation.
They claim big compression ratios: use 100 vision tokens to represent what would be 1,000 text tokens.
This solves the long context problem for LLMs by treating paragraphs as discrete images.
This lets you process larger contexts more efficiently (less compute, fewer tokens), possibly enabling more efficient inference.
In short, if their claims are accurate, this technique would reduce the compute burden for large-context tasks. In other words, you could have a powerful model on your phone, your tablet, or anywhere else. Intelligence becomes ambient. Imagine that all documents and all tools you use are imbued with an ambient artificial intelligence. Depending on your priors this is either thrilling or frightening.
And, in this world, where inference moves away from large, centralized, industrial AI data centers, to edge devices, documents, and tools, aren’t all those large, centralized, industrial AI data centers stranded assets? They no longer need to serve you inference when you ask your friendly chatbot what kind of dog you should get or how you should handle your Aunt Sally’s complaints.
