In an article published a few weeks ago, I argued that the failure of enterprise AI was not due to lack of enthusiasm, adoption challenges, or even model capability. The core issue was architectural: large language models were never built to run a company.

Companies operate on memory, context, feedback, and constraints, while LLMs remain, at their core, systems for predicting text. In a follow-up piece, I proposed that the solution wasn’t simply "better prompts," but a fundamental shift: from tools to systems, from answers to outcomes, from copilots to systems of action, and from prompts to constraints. Enterprise AI cannot be session-based—it must remember.

That argument now requires a third step. Something important is beginning to happen: the systems that are starting to deliver value in enterprise AI don’t resemble better chatbots, copilots, or even improved prompt chains. They look entirely different. And the evidence is already visible.

The Shift from Tools to Systems Is No Longer Theoretical

For the past two years, the AI industry has focused on optimizing the visible layer: larger models, sleeker interfaces, more polished copilots, and increasingly ambitious agents. Yet the clearest signals of value aren’t coming solely from this visible layer. They’re emerging from organizations that are redesigning workflows, embedding AI into core processes, and treating intelligence not as a tool but as infrastructure.

McKinsey’s latest global survey delivers a clear message: AI adoption is widespread, but most organizations have not yet embedded it deeply enough into workflows and processes to generate meaningful enterprise-level benefits. The survey highlights that workflow redesign is one of the strongest contributors to measurable business impact.

This confirms a key point from my earlier articles: the problem wasn’t whether models could answer questions well. The problem was where we were placing them. The organizations making real progress aren’t just "using more AI"—they’re redesigning the company around it.

The Systems That Work Don’t Start from Prompts

The most promising enterprise AI systems emerging today don’t begin with a prompt in the traditional sense. They begin with context: persistent, structured, and governed context.

Anthropic’s engineering team now describes context engineering as the natural evolution beyond prompt engineering. They argue that the real challenge is no longer just how to phrase instructions, but how to manage the entire context state around the model—including system instructions, tools, external data, message history, and environment.

This represents a profound shift. It moves the center of gravity from "What should I ask the model?" to "What environment, state, and constraints should the system already know before any question is asked?"

Anthropic reinforces this perspective in its guidance for long-running agents, emphasizing the need for environment management and pre-configuring agents with the context they require to function effectively across multiple sessions and extended timeframes.

This evolution brings us closer to the vision outlined in my previous two articles: AI success in the enterprise depends not on better models or prompts, but on building systems that operate with deep, persistent context and structural integration into business processes.