What looks like a simple data housekeeping issue – multiple records about the same supplier, in many systems – will have a big impact across your entire P2P operation.
When we unearthed the same supplier listed over 100 times in one organisation’s ERP, AP automation, and procurement systems, it wasn’t just a data quality problem. It was a gateway to millions in lost negotiating power, overpayments, and operational inefficiency. The financial stakes are high, and the time to act is now.
This isn’t an isolated incident. It’s a pervasive problem plaguing every Procure to Pay system architecture, in every organisation. The sooner we acknowledge this, the sooner we can start working towards a solution.
Consider this scenario: A major global supplier had over 20 different codes across different ERP instances within a single national buyer organisation. A pattern that repeated across multiple countries. The result?
In one particularly painful example, we found a single supplier with four different accounts across sites and ERP systems for the same buyer entity. Invoices were loaded on one account. Credit notes on another.
The damage over 12 months:
This wasn’t fraud. It wasn’t incompetence. It was the inevitable consequence of a fragmented data architecture.
If fragmentation is bad in a steady state organisation, it becomes catastrophic during M&A:
The day-one disaster:
The hidden integration costs: Post-merger integration often costs $10-50M, of which 60% is focused on “data harmonisation.” This problem is “solved” using armies of consultants to manually map the data instead of semantic matching. What should take months takes years. What should cost millions costs tens of millions.
To successfully deploy AI in the Procure to Pay space, we’ve built tools that produce unified, clean data. Without this foundation, the promise of AI-powered procurement remains just that – a promise.
The reality for most enterprises: AI initiatives stall not because the technology isn’t ready, but because the underlying data is too fragmented to support even basic automation. Models trained on duplicate suppliers and inconsistent data produce unreliable results that no CFO would trust for critical decisions.
The gap between organisations with harmonised P2P data and those without it isn’t just about efficiency anymore – it’s about competing in an AI-driven future. Every day spent manually reconciling data is another day lost for building the intelligent automation defining procurement leadership in the next decade.
It’s about having the fundamental capability to leverage AI when it matters. And increasingly, it matters for everything from risk prediction to payment optimisation to strategic sourcing.
Traditional approaches to fixing P2P data fragmentation have failed because they treat the symptoms, not the cause. The fundamental issue is that P2P data is inherently about connections – between systems and between the contextual information about each supplier’s contracts, invoices, payments, entities, and hierarchies. These relationships need a new type of architecture designed specifically for connected data.
Before any matching or analysis can begin, data from disparate systems should be mapped to a common schema by:
This harmonisation layer acts as your universal translator, ensuring that downstream processes work with consistent, well-structured data regardless of source system variations.
Traditional matching looks for exact text matches. Semantic matching understands meaning; for example, “IBM UK Ltd”, “International Business Machines (GB)”, and “IBM UK” should be recognised as the same entity. This is achieved by:
Comparing every record to every other record would require billions of computations. Blocking makes it manageable by grouping records into smart “blocks” of candidates which share, for example, the same postcode or similar names. This helps because it:
Once matched, entities live in a connected graph structure:
Now you can instantly answer: “Which suppliers are high-risk, concentrated across business units, and tied to late payments?”
These steps represent a fundamental shift in how enterprises should approach P2P data, from managing records to understanding relationships.
Every day you delay addressing P2P data fragmentation, you’re:
The question isn’t whether to fix your P2P data foundation. It’s whether you’ll do it before these hidden costs become visible crises.
That number you’re looking at? It’s not an IT problem. It’s your business case for transformation.