Stop Investing in AI Until the Content is Fixed

April 30, 2026
by
Michael Kunzler II

There is a pattern in enterprise AI deployments that is consistent enough to be worth naming directly. An organization identifies a compelling use case, evaluates platforms, selects a vendor, and begins implementation. The deployment goes live. Early results are underwhelming, or worse, quietly wrong. The outputs are plausible but unreliable. The team that championed the investment is now managing a confidence problem instead of a performance story.

AI
Strategy

AI Retrieves What You Give It

The diagnosis that follows the standard procurement process usually focuses on the technology. The model is not sophisticated enough. The platform was oversold. The vendor's implementation support was thin. Some of that may be true. AI systems have real limitations, and not every platform delivers what its demos suggest.

But in most of the deployment failures we see, the technology is not the primary variable. The content environment is, as this is the information you are providing the AI system to operate within. That distinction matters because it points to a different intervention, one that has to happen before the next AI investment is made, not during, and certainly not after.

What the Technology Actually Requires

Generative AI in an enterprise context almost always operates through a retrieval layer. The system pulls content from a defined knowledge environment and passes it to a language model as context for a response. That architecture, commonly called RAG (Retrieval-Augmented Generation), is what connects the AI's language capability to an organization's specific content: its products, policies, services, documentation.

The language model brings real capability to that exchange. It can synthesize, summarize, respond conversationally, and operate at a speed and scale no purely human review process can match. It also has real limitations. It cannot reason about whether what it retrieved is current. It cannot flag that a policy was updated six months ago and the version in the knowledge base predates the change. It cannot distinguish between a product description that is authoritative and one that was last edited by someone who has since left the organization. Unless these things are explicity designed into the workflow, it will "hallucinate" a variety of incorrect information.

The model retrieves what is there and generates from what it retrieves, with consistent confidence regardless of the underlying accuracy. That is not a flaw unique to any one platform. It is a structural characteristic of how these systems work. And it means the quality of what the system produces is, to a significant degree, a function of the quality of what it retrieves.

When the content environment is ungoverned, the AI does not produce careful or hedged outputs that signal uncertainty. It produces fluent, confident responses that reflect whatever the environment contains, including outdated information, conflicting versions, and regulatory content that was never structured for retrieval in the first place.

Where Most Organizations Are When They Start

The content environments that enterprise AI systems are deployed against are rarely in the condition those systems require for effective integration. Most digital content environments evolve over time, through accumulated decisions, multiple CMS migrations, shifting ownership, and governance frameworks that were designed as policy rather than workflow.

What that typically produces is an environment where:

Content is high volume but has low structural consistency. Pages were built to render correctly in a browser, not to be retrieved predictably by a machine.

Ownership is defined at the department level, not the content type level. Marketing owns the website in the same way a building manager owns a lobby: they are responsible for it in aggregate but not always accountable for the accuracy of any specific thing inside it.

Governance exists as intention rather than infrastructure. There is usually a process for publishing new content. There is rarely a reliable mechanism for identifying when existing content has become inaccurate, who is responsible for updating it, or whether an update has propagated consistently across all the places that content appears.

Deploying AI against that environment does not produce intelligent outputs. It produces fast outputs that inherit all of the environment's existing problems, at a scale and speed that makes those problems significantly harder to detect and correct.

The Investment Sequence That Produces This Outcome

The pattern is not primarily a vendor problem or a technology problem. It is a sequencing problem.

AI investments are evaluated and approved based on the capability the platform demonstrates. The governance and structural conditions required for that capability to perform are treated as implementation details, addressed after the contract is signed or, more commonly, discovered as problems after the deployment is live.

The result is that organizations end up in a position where the AI investment is real, the performance is disappointing, and the fix requires exactly the kind of foundational content work that should have preceded the investment. That work is now more expensive, more politically complicated, and harder to fund because the budget was spent on the platform.

The question that gets skipped in most procurement cycles is whether the content environment the AI will retrieve from is in a condition that can support the deployment. Not whether the organization is excited about AI. Not whether the use case is compelling. Whether the specific content domain that will feed the retrieval layer is structured, governed, owned at the right level of specificity, and current enough to be trusted.

What an Improved Sequence Looks Like

The organizations that get more from AI investments are not necessarily the ones with better technology, rather they are the ones whose content environments were ready before deployment began, or who scoped the initial deployment to the portion of their environment that was.

That sequencing starts with a content environment assessment: a structured audit of the domains the AI will retrieve from, evaluated against the conditions that make retrieval reliable. What is structured and what is not. Where ownership is defined clearly enough to be actionable and where it is not. Where governance is a workflow and where it is a policy document that nobody enforces. What is current and what has accumulated without a review process.

That assessment produces two things. First, a realistic picture of what can be deployed against today without significant remediation. Second, a prioritized list of what needs to be addressed before scope expands.

The initial deployment scope follows from that picture. Not from what the platform can technically do, but from what the content environment can reliably support. A narrow, well-scoped deployment against a governed content domain produces better results than a broad deployment against a fragmented one, builds organizational confidence in the capability, and creates the evidence base required to make the case for expanding scope.

Starting small is not a hedge against ambition. It is the necessary sequencing that makes ambition sustainable.

The Practical Implication for CIOs and CMOs

The decision that determines most AI deployment outcomes is not which platform to select. Instead, the primary determination emerges from a decision to assess the content environment before the platform is selected.

That assessment is not a delay. It is what makes the platform decision defensible, because it means the deployment scope is defined by what the environment can actually support rather than by what the demo suggested was possible.

For organizations that have already made the investment and are managing a performance gap, the same logic applies in reverse. Before adjusting the platform configuration, expanding the scope, or switching vendors, it is worth asking whether the content environment is the variable that needs attention. In most cases, it is at least part of the issue, if not most of it.

The technology will continue to improve. The models will get better at reasoning about uncertainty, at flagging retrieval confidence, at operating against less structured environments. Those improvements are real and they matter, but they do not change the fundamental dependency: a retrieval system operates against what the organization gives it. Improving the model's capability does not make ungoverned content more reliable. It will simply make the output more persuasive, which, in a poorly governed environment, is a more sophisticated version of the same problem.

The content environment is the variable most organizations can actually control. It is also the one most organizations have not assessed before they write the check.

Get monthly insights on building smarter, more effective digital experiences—straight from the team at C2.