In boardrooms and technology war rooms alike, a deceptively simple question is reshaping how organisations think about artificial intelligence, cloud migration, and digital transformation: whose data is it? The answer and the infrastructure built around it is fast becoming the single most consequential factor in whether an AI initiative thrives or collapses.
By: Norman Roberts, Chief Technology Officer at Prescient Investment Management.
As we move through 2026, the evidence is unambiguous. The visibility and lineage of a company's data is not a nice-to-have governance exercise. It is the foundation upon which every AI model, every automated insight, and every integrated platform is built. Without it, even the most sophisticated technology becomes an expensive gamble.
The Failure Epidemic: What the Numbers Say
The scale of AI project failure is staggering, and the primary culprit is consistently the same: organisations do not understand their own data well enough to trust it with critical decisions.
80% of AI projects fail to reach meaningful production — twice the failure rate of non-AI IT projects. - Rand Corporation, 2024
42% of companies abandoned most of their AI initiatives in 2025, up sharply from just 17% in 2024. - S&P Global Market Intelligence, 2025
95% of corporate AI pilots produce zero measurable bottom-line impact despite $30–40 billion in enterprise investment. - MIT GENAI Divide Report, 2025
60% of AI projects unsupported by AI-ready data will be abandoned by the end of 2026. - Gartner, 2025-2026 Forecast
43% of data leaders cite data quality and readiness as the top obstacle to AI success. - Informatica CDO Insights, 2025
A 2024 Forrester Research survey of 500 enterprise data leaders found that 73% identified data quality and completeness as the primary barrier to AI success, ranking it above model accuracy, compute costs, and talent shortages combined. These are not isolated findings. They are a systemic signal.
The Root Cause: You Cannot Govern What You Cannot See
The failure pattern is remarkably consistent across industries and company sizes. Organisations invest in AI tooling, build proof-of-concepts that impress in demos, and then watch adoption stall or outputs become unreliable at scale. What breaks is rarely the model. What breaks is the invisible infrastructure around it, the data pipelines, the ownership boundaries, the lineage gaps, and the absence of enforceable policies on what data can be consumed, by whom, and for what purpose.
Gartner's 2024 analysis found that only 48% of AI projects make it into production, and those that do take an average of eight months from prototype to deployment. The bottleneck is almost always data readiness not model sophistication. A well-trained model fed corrupt, incomplete, or unauthorised data will hallucinate. It will produce confident answers that are factually wrong. In a regulated industry such as asset management or financial services, that is not merely a technical embarrassment: it is a compliance event.
What Modernisation Actually Looks Like
Genuine data modernisation is not about adopting the newest tool or migrating workloads to the cloud for the sake of it. It is a disciplined, architectural commitment to knowing exactly where every piece of data comes from, who owns it, what transformations it has undergone, and what governance rules constrain its use. For decision makers, this means investing disproportionately in the foundations before building the penthouse.
The organisations consistently outperforming their peers share a recognisable pattern. They earmark 50 to 70 percent of their AI timeline and budget for data readiness extraction, normalization, governance metadata, quality dashboards, and retention controls before a single model is trained for production. McKinsey's 2025 AI survey confirms that organisations reporting significant financial returns are twice as likely to have redesigned end-to-end data workflows before selecting their modelling techniques.
Architecturally, the most effective approach is a layered data platform: raw ingestion at the bronze layer, cleansed and validated data at the silver layer, and curated, business-ready datasets at the gold layer. This Medallion Architecture, increasingly adopted by data-mature organisations, creates natural checkpoints for lineage tracking, data quality validation, and ownership assignment at every stage of the data journey.
Prescient Investment Management: Modernisation in Practice
At Prescient Investment Management, the journey reflects precisely these principles. PIM is not starting from zero the firm operates with a well-established data environment built over years of institutional discipline. But as AI initiatives mature and integration surfaces expand, the team recognised that 'established' is not the same as 'governed for the AI era'.
PIM's current modernisation programme is structured around three pillars. The first is a cloud-native data foundation built on AWS, implementing a Medallion Architecture with OpenMetadata governance ensuring that every dataset consumed by any downstream platform or AI service has a traceable, audited lineage from source to insight. The second is CRM integration and workflow automation, where clean, authorised data flows between systems without manual reconciliation or silent data drift. The third is AI governance itself: a formalised use-case register, tool justification frameworks, and clear policies on which models can consume which datasets under what conditions.
The motivation is direct: without knowing exactly what data feeds a model, the outputs of that model cannot be trusted. Worse, in an asset management context where calculations inform client reporting, portfolio construction, or regulatory submissions, unverified data lineage creates real liability. The work of understanding ownership, implementing guardrails, and building observability into every data pipeline is the work that makes AI outputs defensible not merely plausible.
For Decision Makers: The Questions That Matter
If you are weighing an AI or modernisation investment in 2026, the technology decision is secondary. The primary questions are organisational and architectural:
- Can you trace every data asset your AI systems consume back to an authoritative, owned source?
- Do you have enforced policies, not just documented ones, that prevent unauthorised data from entering critical pipelines?
- Is your data quality monitored continuously, with observable alerts before failures reach production?
- Are data ownership responsibilities assigned to named individuals not departments or systems?
- Does your AI governance framework address hallucination risk specifically and what data controls are in place to reduce it?
If you cannot answer yes to each of these, your data is not yet ready to be the foundation of an AI platform. And a platform built on an uncertain foundation will, with statistical inevitability, join the majority that fail.
The Competitive Advantage of Knowing
The upside is equally compelling. A 2024 ZoomInfo analysis of Fortune 500 companies found that organisations leveraging advanced data intelligence platforms achieved five times the revenue growth, 89% higher profits, and 2.5 times higher valuations compared to industry peers. The data advantage compounds. Companies that understand their data estate today are not just better placed for their next AI project — they are building a structural moat that becomes harder for competitors to close over time.
The question, then, is not whether your organisation can afford to invest in data ownership, lineage, and governance. The question is whether it can afford not to. In 2026, whose data it is and whether you can prove it is the difference between AI that delivers and AI that disappoints.
Disclaimer
Prescient Investment Management (Pty) Ltd is an authorised Financial Services Provider (FSP 612).
The information contained herein is provided for general information purposes only. The information and does not constitute or form part of any offer to issue or sell or any solicitation of any offer to subscribe for or purchase any particular investments. Opinions and views expressed in this document may be changed without notice at any time after publication and are, unless otherwise stated, those of the author and all rights are reserved. The information contained herein may contain proprietary information. The content of any document released or posted by Prescient is for information purposes only and is protected by copy right laws. We therefore disclaim any liability for any loss, liability, damage (whether direct or consequential) or expense of any nature whatsoever which may be suffered as a result of or which may be attributable directly or indirectly to the use of or reliance upon the information. For more information, visit www.prescient.co.zat www.prescient.co.za