justin․searls․co

Models are commodities, harnesses are differentiators

Ben Thompson's latest rests on a single load-bearing assumption: that the harness and the model are tightly coupled, the way Apple's hardware and software are.

It follows, then, that if agents require integration between model and harness, that the companies building that integration—specifically Anthropic and OpenAI—are actually poised to be significantly more profitable than it might have seemed as recently as late last year.

This assumption is incorrect. To date, agents have not clearly benefited from proprietary integration to their favored model. I thought this was obvious, but I sometimes forget that not everybody else has made it their full-time hobby to mix and match various coding agents with various large language models for the past year.

Here's how I see it.

The API surface of a frontier model is text in, text out. Images in, images out. Audio in, audio out. AI models are among the most immediately pluggable and therefore commoditizable innovations in the history of software—right up there with UNIX pipes and their simple promise of text in, text out. Countless forks and TUIs have already demonstrated that swapping the underlying model is a marginal concern, not a structural one. The harness doesn't integrate with the model. It integrates with the user's world—files, tools, workflows, intent—and calls the model as one resource among many.

This distinction matters enormously. A high-quality harness paired with a mediocre model can accomplish incredible things. A frontier model paired with a poorly-considered harness can barely get out of the gate. If model-harness integration were truly the differentiator, the model would indeed be irreplaceable and invaluable. But in practice, models have turned out to be the most replaceable part of the stack. And every recent advance in tool use, verification, and orchestration has only made this clearer—the harness gets more valuable while the specific model behind it gets more interchangeable. As long as multiple companies are credibly in the frontier model game, there will be cutthroat competition and relentless downward pressure on price. And while it may not happen soon, models and hardware will eventually advance to the point that a local open model is good enough—and then the frontier labs won't just be commoditized, they'll be cut out entirely.

Thompson brings up Microsoft's new Copilot Cowork bundling initiative as evidence that model-harness integration is where value accrues:

Fast forward to last week, however, when Microsoft revealed how they will handle the potential business impact of AI reducing seats, which is a bit of a problem for their seat-based business model: the company is going to bundle AI into a new higher-tiered enterprise offering, E7, which is going to cost twice as much — $99 per seat per month — as the formerly top-of-the-line E5. That's a big increase, which Microsoft needs to justify with AI that actually makes those seats more productive, and the product they launched with the new bundle was Copilot Cowork.

Microsoft is indeed in a tight spot, but not because the model makers have a monopoly on harnesses. Instead, blame their outlandish investments in somebody else's model, their piss-poor track record for delivering a single worthwhile Copilot-branded piece of software outside the GitHub org, and their over-reliance on per-seat licensing that is existentially at odds with a strategy that, if successful, would drastically reduce the number of seats its customers need. None of that tells us anything about whether models and harnesses are structurally coupled. It only tells us what we already know: that Microsoft has put itself in a strategically weak position—a strategy that presupposed they would somehow stop sucking at building software people actually want to use.

Apple's situation is a cleaner test of the thesis. Dediu is probably right that Apple has a tremendous opportunity precisely because they can build a harness and swap the underlying model. Will they build an incredible harness? No—they'll build one that's just good enough to keep customers buying their hardware. But the strategic logic is sound: focus on the human interface and the real-world problems it can solve, and don't cede the narrative to implementation details like which model is running underneath. The truly great harnesses will probably come from smaller, more nimble startups who understand that the opportunity isn't in training the next frontier model—it's in connecting the horsepower of any capable model to the thousand different domains waiting to be served.

The frontier labs attracted an outsized share of attention and funding because OpenAI happened to hit one out of the park with ChatGPT as a product, and this conflation continues to produce analyses that make the same mistake Thompson is making. ChatGPT went so mainstream that it's treated as a blockbuster product, but I suspect we'll look back on it merely as a (historically impactful) proof of concept. The work of building a chatbot and the work of training a frontier model are only incidentally related. The thousands of chatbot systems that have sprung up since ChatGPT's release should be all the evidence we need that building a great chatbot never required building a great model—OpenAI just happened to do both.

At the end of the day, the utility of models is owed to the fact that they are black boxes with simple interfaces. Text in, text out. Audio in, audio out. Increasingly, video in, video out. But that simplicity is also what makes them commodities. And leads don't last long enough for anyone to establish a moat—the underlying research unlocks improvements in the open, and ubiquitous access to everyone's models facilitates distillation of competing ones.

Harnesses couldn't be more different. Harnesses are the gooey, ever-shifting integration of human-computer interface. They connect people's needs to implementations designed to meet them. With a little cleverness and elbow grease, they can be dispatched to verify that those implementations work. They can be assigned to orchestrate other agents across systems in separate geographies, and over sessions conducted at different times.

The model is the engine. The harness is the car. And history has shown—over and over—that you don't need to manufacture your own engine to build a great car.


Got a taste for hot, fresh takes?

Then you're in luck, because you'll pay $0 for my 2¢ when you subscribe to my work, whether via RSS or your favorite social network.

I also have a monthly newsletter where I write high-tempo, thought-provoking essays about life, in case that's more your speed:

And if you'd rather give your eyes a rest and your ears a workout, might I suggest my long-form solo podcast, Breaking Change? Odds are, you haven't heard anything quite like it.