Most AI projects succeed or fail based on the data behind them. Here's what that means - and what you can do about it.
Companies debate the right AI model, the best provider, the perfect strategy. They rarely debate their data. Yet data quality is the most common reason why mid-market AI projects fail to deliver on their promise.
The principle isn't new: garbage in, garbage out. A language model working with inconsistent, incomplete or incorrect data will produce inconsistent, incomplete or incorrect results - reliably and at high speed.
Poor data isn't always obviously wrong. More often it's incomplete (fields left empty because they weren't needed day-to-day), inconsistent (the same date in three different formats), outdated (master data that hasn't been maintained for two years) or isolated (critical information sitting in emails or PDFs, not in structured fields).
For AI automation, that last category is particularly critical. When relevant information is trapped in free text, handwritten notes or scanned documents, the first step isn't AI - it's structuring.
Silos: Sales has CRM data, accounting has ERP data, project management has Excel files. All three contain relevant customer data - none is complete. An AI system designed to work across departments fails at the boundaries of these silos.
Historical inconsistency: Three years ago, the CRM system was changed. Data migration was incomplete. Legacy data was mixed with new formats. Anyone building on this data today is building on shaky foundations.
Lack of input discipline: Mandatory fields get bypassed, free text is used where structured fields should be, abbreviations aren't standardised. This isn't staff failure - it's a structural problem that tools and processes must solve together.
A realistic AI audit always looks at the data first. Not to slow things down, but to assess what the first sensible step should be. Sometimes that's AI implementation. Sometimes it's a data cleansing project first - which then creates the foundation for far more effective AI.
Companies that invest directly in AI technology without understanding their data are buying a powerful tool for poorly prepared work. The tool isn't the problem.
The pragmatic approach: choose a single process earmarked for AI automation and analyse the data situation specifically for that. What input data does the planned system need? Is this available? In what format and quality?
This analysis reveals whether cleansing is needed, whether capture processes need adjusting, or whether the data situation is already sufficient for an initial pilot. This focused approach is far more efficient than company-wide data cleansing without a concrete use case.
This sounds sobering, but it isn't. The good news: almost every company has better data than it thinks - it just doesn't always sit where it's needed. Taking a structured look at your data situation before deploying AI technology is the crucial difference between a pilot that works and one that ends up in a drawer after six weeks.
In an AI audit, I look at your data situation first - honestly and without any sales agenda.