
In Post 1, I argued that machine readability is a meaning problem, not a format problem. In Post 2, I showed that solving it makes things better for the humans in your operation - your subject matter experts, your end users, and anyone who has ever spent time reconciling product data across disconnected systems.
In this post (3 of 9), I want to turn to a question that will separate organizations that use AI well from those running remediation programmes in three years:
How do we operationalize AI for the next 3+ years - across multiple tools, providers, and use cases?
Specifically, I want to make the case that the information backbone is not just useful for AI. It is what makes AI operationally viable - because it is what gives you the freedom to use AI on your own terms: to swap tools, try new providers, and move on from them when something better comes along, without it becoming a programme of work every time.
Bottom line up front:
AI systems do not generate reliable outputs from unreliable foundations. If the information they consume lacks governed meaning, every output is a guess. But organizations that build a proper information backbone before layering AI on top of it gain something beyond better outputs: they gain the freedom to swap AI tools, change providers, and adopt new capabilities - without touching their information systems.
And before you scroll past - this isn't about hallucinations, and it isn't a privacy lecture about keeping your data to yourself. Both of those conversations are done. This is about something more foundational and more durable: what your organization's information needs to look like so that AI serves you, rather than the other way around.
The pressure to deploy AI is real and it is coming from every direction - boards, technology teams, competitors, analysts. The promise is significant: faster processing, better insights, automated reporting, natural language Q&A, more capable supply chain systems, the list goes on.
But underneath the pressure is a question that is rarely asked clearly enough: What is the AI actually consuming? When an AI agent makes a decision about what information to display, gives a machine instructions for a contextual product use, processes a regulatory query, summarises a product record, or flags a substance restriction - where is it drawing its understanding from? Is it reasoning from governed, verified meaning? Or is it assembling an answer from whatever it can find, at the moment it is asked, to give a best guess?
That distinction matters enormously in regulated and high trust environments. And it is the distinction that the information backbone resolves.
The default mode for most AI deployments - particularly those built on large language models - is to assemble meaning at the point of query. The AI receives a question, searches its training data or available context, and constructs a response that is plausible given what it has seen.
This works well for general knowledge. It works poorly for regulated, organization-specific information where precision is not optional.
If your product data contains inconsistent classifications, ambiguous field names, or ungoverned reference data, the AI will encounter that ambiguity at query time and do its best to resolve it. Sometimes it will resolve it correctly. Sometimes it will not. And critically - you will often not know which is which until something goes wrong.
This is not a failure of the AI. It is a failure of the foundation the AI is working from. The AI is doing exactly what it is designed to do: making its best response from the available information. The problem is that "best response" is not good enough in environments where outputs need to be auditable, reproducible, and defensible.
An information backbone changes what AI has access to - fundamentally. Instead of assembling meaning from scattered, inconsistently defined sources, the AI consumes from a model where meaning is already explicit, governed, and verified.
A product's composition isn't just a flat list of data properties to be interpreted - it is a structured set of relationships: components, materials, substances, assertions, regulatory classifications, each connected to the others in ways that have been defined, reviewed, and approved. A substance code isn't an opaque identifier that the AI has to guess the significance of - it is a governed reference that the backbone already maps to its regulatory context.
When AI operates on top of that foundation, "best guessing" becomes unnecessary for the things that matter most. The backbone already holds the answer. The AI doesn't need to guess whether a substance is restricted - the relationship is explicit. It doesn't need to interpret what a classification means - the governed model defines it. The measurement units do not need to be guessed from a set of unstructured information - the units of measure are explicit.
The room for error is not just reduced. For the core facts of your information model, the risk of error is eliminated.
Not only is the information backbone-based approach explicit and trustworthy for those reasons, it is also more efficient because the scope of the information required is known and clear from the question scope, so the AI tools do not need to include extraneous information.
This is what it means to ground AI in absolute surety. Not prompting it more carefully. Not adding retrieval layers on top of unstructured data. Building the governed semantic foundation first, and letting the AI consume from it.
A technically informed response to this argument is that fine-tuning or retrieval-augmented generation (RAG) can close the accuracy gap - you train the model on your own data, or you give it access to your documents at query time, and it performs better than a general-purpose model would.
This is true. Fine-tuning and RAG do improve accuracy. For many use cases, that improvement is sufficient. If you are building a customer-facing assistant that answers general product questions, or a tool that helps internal teams navigate documentation, "significantly better than baseline" may be entirely acceptable. Some error rate is tolerable. A wrong answer occasionally is a recoverable situation.
But in regulatory and high-trust contexts, "significantly better" is not the standard. The standard is correct. Every time.
A regulatory submission that contains a misclassified substance because the AI resolved an ambiguity incorrectly is not a minor quality issue - it is a compliance failure. A product passport that maps a material to the wrong regulatory code because the AI drew on an inconsistently governed data source is not an acceptable margin of error - it is a record that cannot be defended under audit. The question is not whether fine-tuning or RAG produces good results on average. It is whether it produces verifiable, traceable, auditable results in every instance that matters.
The deeper problem is that fine-tuning and RAG on poorly governed data inherits the ambiguity of that data. If your classifications are inconsistent, fine-tuning learns those inconsistencies. If your reference data contains multiple versions of the same concept, RAG retrieves whichever version it finds most relevant at query time - which may not be the governed one. Improving the AI layer does not fix the foundation. It papers over it, with a degree of sophistication that makes the underlying problem harder to see.
The information backbone is not an alternative to RAG or fine-tuning, it is what makes any AI approach more reliable and efficient. When AI retrieves from a governed model - where relationships are explicit, classifications are controlled, and reference data is maintained - retrieval is accurate because the source and scope are accurate. The AI is not resolving ambiguity, it is reading verified, scoped meaning. That is a fundamentally different proposition, and in regulated environments, it is the only proposition that holds.
There is a structural reason why this works - and it is worth understanding, because it has implications that stretch well beyond any single AI deployment.
A proper information backbone built on FAIR data principles uses open, interoperable standards to express meaning: entities, relationships, classifications, and reference data in a form that any compliant system can read and reason about. This is not proprietary to any AI tool or vendor. It is not tied to a particular model's training data or a specific platform's interpretation.
This matters because the backbone's meaning is legible to any machine that adheres to those standards - not just today's AI tools, but whatever comes next. The governed relationships in the backbone do not need to be re-explained to a new system. They are expressed in a language that any standards-compliant tool can consume directly. The meaning is expressed independently of any tool's proprietary format - so onboarding a new AI system is a connection, not a migration.
This is the operational point that matters most for a COO making decisions now, and it is the one most often missed in AI strategy conversations.
The AI market is moving fast, and it will keep moving. The AI tool that represents best value today may not be the best option in eighteen months. New models emerge. Pricing changes. Capabilities shift. A provider that makes sense now may be the wrong choice in two years - for commercial reasons, performance reasons, or regulatory ones.
The organizations that will navigate this well are not the ones that have made the deepest commitment to a single AI vendor. They are the ones that have kept their options open by building the stable foundation that any vendor can serve.
If your AI tools are consuming from a governed information backbone, switching providers is a configuration change, not an operational crisis. The new tool connects to the same backbone. It reads the same governed meaning. It produces outputs from the same verified foundation. Your information systems do not change. Your data does not need to be re-mapped. Your subject matter experts do not need to re-validate their work for a new platform.
If, on the other hand, your AI tools are deeply entangled with the way your information is currently stored and labelled - if meaning is assembled at query time from a particular system's particular structure - then switching providers means unpicking that entanglement. It means rebuilding the context the AI relies on. It means, in practice, a programme of work every time you want to change tools.
The backbone gives you optionality. The absence of a backbone turns every AI decision into a long-term commitment you never intended to make.
The organizations that will use AI most effectively over the next 3+ years are not necessarily the ones deploying it most aggressively today.
They are the ones building the foundation that makes AI outputs trustworthy - and that keeps AI tools interchangeable.
That foundation is the information backbone. Build it with open standards, and govern the meaning it holds.
Getting the foundation right now pays forward - you can let the AI tools come and go on top of it.
Post 4: What do we mean by "information backbone"? - a plain-language definition for operational leaders, and why it matters that your organization understands what it is before deciding how to build it.
Post 5: The quiet power of reference data - the unglamorous work that gives your backbone a stable vocabulary, and why getting it wrong undermines everything else.
Post 1: Preparing for true machine-readable digital product labels - Machine readability is a meaning problem, not a format problem. Most organizations focus on file formats and miss the foundational architecture problem entirely. This is what it actually demands from your organisation.
Post 2: Machine-readable isn't just for machines - Better information architecture foundations improve the experience of the humans who work with product data every day - your SMEs, compliance teams, and end users. Better foundations for machines are better foundations for people.