From Hallucinations to Trust: Context Engineering for Enterprise AI
Context engineering is the systematic practice of designing and controlling the information AI models consume at runtime, ensuring outputs are accurate, auditable, and compliant.
But the debate is missing a more fundamental question: what do captured decision traces actually enable?
Context graphs will make AI dramatically better at reasoning. But they won’t teach it judgment. And in business, asymmetric value comes from judgment, not reasoning. We’ve seen this movie before with self-driving. Despite comprehensive decision traces and massive amounts of data it has been incredibly hard to bridge the gap from reasoning to judgment.
Let’s not make the mistake of assuming that more context will solve a judgment problems with reasoning.
A context graph is more than a log of data. It’s a record of the reasoning behind decisions. When a renewal agent proposes a 20% discount despite a 10% policy cap, the context graph captures not just the discount, but the approval chain, the incident history from PagerDuty, the escalation threads from Zendesk, and the precedent from prior exceptions.
The thesis: if we capture these decision traces comprehensively, AI agents will learn to navigate enterprise decision-making the same way they learned to write code and generate content.
There’s just one problem.
The context graphs movement rests on a powerful parallel:
What happened with LLMs and knowledge:
The implicit assumption about context graphs:
This assumes business decisions are primarily reasoning problems. They’re not. They’re judgment problems.
Reasoning problems are deterministic given sufficient context. Calculate optimal inventory levels, route support tickets, generate sales forecasts, approve expenses within policy. Given complete information, there’s a “correct” answer. Context graphs will genuinely help here.
Judgment problems involve weighing incommensurable values under uncertainty. The same inputs can demand opposite decisions depending on latent intent, unobservable state, and relationship dynamics that exist outside any system.
You can’t feed reasoning traces to solve judgment problems. It’s a category error.

Autonomous driving is the perfect example for this, with over:
Yet full autonomy has taken so much effort, yet proven hard because driving is as much a judgment problem as rules and reasoning. Business decisions face the same challenges at greater complexity:
| Challenge | Driving | Business |
| Interpreting Intent | Pedestrian at crosswalk: waiting for light or about to jaywalk? The same scene requires different responses. | Customer requests discount: price shopping or genuinely constrained? The same email requires different responses. |
| Reading Unspoken Cues | Blinker on: actually merging or forgot it three turns ago? Watch the wheels, not the signal. | “I strongly suggest we make an exception”: suggestion, directive, or political cover? Message ≠ meaning. |
| Applying Unwritten Rules | Zipper merge, courtesy wave, when to yield all vary by region and situation. This will not be found in any rulebook. | Which issues need escalation? Which exceptions need paper trails? Official process will always differ from actual practice based on situations. |
| Asymmetric Risk | Hesitating costs seconds; misjudging costs lives. Context determines acceptable risk. | Losing a $10K deal costs $10K; damaging a relationship could cost millions. Automation can’t calculate this. |
These examples demonstrate that the same data point demands different actions depending on context outside any system.
This is why Arvind Jain is right: “You can’t capture the why, only the how.” Even comprehensive process traces miss the judgment layer.
Maybe, but judgment depends on counterfactuals that context graphs cannot capture. When a leader approves a “strategic” exception, the graph records the approval, not the similar decisions she declined, the precedents she avoided setting, or the organizational capital she chose not to spend. Judgment is defined by the roads not taken; context graphs only see the roads traveled. That is why trying to learn judgment from decision traces is fundamentally different from learning reasoning from data.
Context graphs will make AI agents dramatically more capable at reasoning tasks. But decisions that create asymmetric value building companies, developing talent, and navigating crises will remain human for the foreseeable future.
Why? Because, we’re feeding reasoning traces to solve judgment problems.
The limitation isn’t capturing context. It’s the category error: better reasoning enables decision support, not decision replacement.
The opportunity isn’t replacing human judgment. It’s augmenting it.
As CEO of Nexla and creator of Express.dev, I’m deeply invested in building the context engineering layer.
In addition to being improving reasoning, Context Graphs also unlock
The pattern: context graphs don’t replace judgment. They make reasoning scale, coordinate, and validate itself. That’s the real opportunity.
Here’s what we can do in practice:
The companies that win will not replace human judgment. They will design systems that know exactly when to defer to it.
Context engineering is the systematic practice of designing and controlling the information AI models consume at runtime, ensuring outputs are accurate, auditable, and compliant.
In episode four of DatAInnovators & Builders, BigID’s Stephen Gatchell explains the data governance gap blocking AI production, why unstructured data breaks legacy models, and how data product frameworks enable scale.
In the News: betanews.com: In this Q&A, Saket Saurabh explains why context engineering is key to reliable, compliant, and intelligent enterprise AI workflows.