All articles

Brittle Data Has a Cause. Data Malleability Is the Cure.

Brittle Data Has a Cause. Data Malleability Is the Cure.
Data debt names what went wrong. Data Malleability names the capability that prevents it. This article introduces a new framework for building data that absorbs change instead of fracturing under it.
Kunal Sharma
Kunal
Sharma
Vice President, Data Management
View bio

In my last article, “Stop Treating Data Debt Like a Technical Problem,” I made the case that data debt is an organizational failure, not a technical one. The cost accumulates not from bad architecture choices but from deferred decisions about who owns the data, what it means, and whether it can actually be trusted. If an organization cannot name the business owner of its most critical data elements, the next transformation is already carrying debt.

That argument is easy to recognize, but the harder question is what to call the capability that prevents it.

The invoice doesn’t go to the person who caused the problem. It goes to the person closest to the data.

I worked with a large building materials distributor for several years. They had a talented senior data leader who understood the problem better than almost anyone in the organization. He led a program that took the organization from paper reports and highlighters to modern dashboards, a genuine transformation in how the business consumed information. People could see more, faster, than they ever had before. But the data feeding those dashboards still carried the ambiguities nobody had resolved. Ownership was assumed, not assigned. Definitions drifted across departments. The dashboards were clean. What flowed through them was not. And the gap between what he could build and what the business could trust became the ceiling on everything he tried to demonstrate.

When a contractor walked in and bought lumber without a project code, the system had no way to know whether that lumber was destined for a Lennar development or a private side job. That single ambiguous transaction, multiplied across thousands of daily purchases, meant that nobody truly owned the definition of a customer. A configurator project stalled because the underlying product data was not reliable enough to build on. E-commerce could not launch. The senior data leader could see every one of these problems clearly. He could describe, document, and escalate them. What he could not do was force the domains that created the ambiguity to resolve it. Business ownership of data was assumed rather than built. And in its absence, the gap between what the data infrastructure could do and what the business allowed it to do became his permanent performance review.

Then the company acquired a competitor. Within the first quarter of integration, it was clear whose data practices were stronger. It was not the acquirer’s. As the organization worked through the complexity of combining two companies, data became a fault line in the transformation. The senior data leader who had been building and delivering for years found himself in a defensive posture he should never have been in. Governance blockages, data quality problems, and unresolved definitions, none of which he owned, all of which landed on him. The acquisition held up a mirror. What it reflected was not his failure. It was the organizational capability that was never built around him.

Every framework assumes the capability exists. Almost none of them build it.

The architecture was not the problem. Organizations running modern cloud platforms fail at this for the same reason organizations on decade-old systems fail at it. The stack is never the variable. The same dynamic surfaces in financial services, healthcare, manufacturing, and retail. The details change. The underlying failure remains the same.

The pattern repeats because the assumption at the center of every major data framework goes unspoken: that the people responsible for data already know how to think about it semantically, define it without ambiguity, and own it at the domain level. Data Mesh requires business owners who can define and steward data products. Most organizations do not have them. Not because their people are incapable, but because no one ever deliberately built that capability. The framework arrived before the foundation did. And the senior data leader inherits the consequences every time.

Gold doesn’t start as gold. Neither does malleable data.

Data debt, as I defined it previously, is backward-looking. It names what accumulated in the absence of good decisions. But the leaders who recognized their organizations in that diagnosis kept asking the same follow-on question. What was actually missing? What is the capability that, had it existed, would have prevented this?

That capability is Data Malleability.

The word is deliberate. Gold is malleable. It can be shaped precisely, hold that form with integrity, and be reshaped when the work demands it without losing its fundamental nature. But gold does not begin that way. Raw ore undergoes an intensive refinement process. Heat removes the impurities. What survives is something with a different character than what went in, more consistent, more trustworthy, more valuable precisely because of what was burned away.

And pure gold, for all its properties, is actually too soft to hold its shape under sustained pressure. It becomes durable when alloyed with other metals in the right combination. The alloy is what makes it workable at scale. What makes it last?

Data Malleability works the same way. The refinement is the organizational work of clarifying what data means, who owns it, and what trust looks like. The alloy is the combination of capabilities that hold the system together under pressure: working as a single structure rather than separate initiatives. Neither the refinement nor the alloying happens by accident. Both require deliberate construction. The specific composition of that alloy is worth its own conversation, and I will have it in the next article. For now, the principle holds: malleable data, like gold, earns its properties through process and combination. It does not arrive that way.

That property is also why the data industry settled on gold as its highest certification tier in medallion architecture. The question this framework asks is whether your organization has built the capacity to keep data at that standard as the business evolves.

Some will ask why not resilience or agility. Both matter. Resilience describes how much a data environment can absorb before something downstream breaks. Agility is about how fast the organization recovers when something goes wrong. Malleability asks a third question that neither of them addresses: whether data carries enough meaning, ownership, and context that the people and systems relying on it can trust what it says without calling anyone for clarification. Most organizations have never formally asked that question.

Data Malleability is the organizational ability to reshape and rely on data as the business evolves, without the rework and eroded trust that make transformation so expensive. Malleable data absorbs change. Brittle data fractures under it. For organizations running AI at any scale, the stakes are higher still: brittle data does not produce one wrong output; it produces that error across every decision an agent makes before anyone notices.

Governance didn’t fail. It was pointed in the wrong direction.

Governance did not fail; it was oriented toward control rather than capability. Malleability does not replace it; it reorients it. From auditing what exists to keeping data fit for purpose as the business evolves.

“Debt compounds. Malleability is how you overcome it.”

You can’t manage what you haven’t named.

What makes this more than a useful metaphor is that malleability is observable.

How long does it take to restore trusted reporting after an upstream system change, and what does it cost when that window stretches to weeks? How many downstream systems break when a single upstream definition shifts? Can a new team member, or an AI agent, understand what a data asset means from the documentation alone without calling anyone?

These are not abstract concerns. They are measurable signals. Individually, they point to specific weaknesses. Taken together, they describe something broader: how well an organization’s data can adapt to change without losing meaning, trust, or usability. Most organizations do not track these signals, and as a result, they cannot explain why change is so expensive or why trust erodes over time.

That distributor I described earlier is still working through it. They are years into an ERP migration, and in their most recent annual filing, they named data governance as a risk to the implementation itself. That kind of transparency takes courage. It means leadership understands what the real problem is. But understanding the problem and having built the capability to solve it are different things. The work of making data malleable does not happen during a migration. It has to come before one.

The organizations that do track these signals have built something their competitors have not: data that can evolve without breaking. They are not guessing whether their data can support change. They can measure it.

That is what data malleability looks like in practice.

That measurement framework is what I am introducing next.

Other articles

The $800,000 Problem Hiding in Your Analytics Team

The $800,000 Problem Hiding in Your Analytics Team

Data Governance
Best Practices
Case Study
When 40% of your analytics team’s time goes toward hunting for data instead of analyzing it, the cost adds up fast. For one healthcare system, that hidden waste reached $800,000 a year. Here’s what actually fixed it.
Building a Data Quality Foundation for a Community College Using an AI-powered DQ Platform

Building a Data Quality Foundation for a Community College Using an AI-powered DQ Platform

Case Study
Data Governance
Best Practices
Following a Student Information System migration, a mid-sized community college faced persistent data reliability issues. A structured data quality program using an AI-powered platform established a scalable capability within six months.
Don’t Slow Down the AI Train—Fix the Tracks

Don’t Slow Down the AI Train—Fix the Tracks

Data Governance
Best Practices
Data Value Realization
A response to Tom Davenport’s call to slow AI development. The problem isn’t that AI is too fast—it’s that organizations are too unprepared. The answer is better foundations, not less innovation.
Client testimonial
The Definian team was great to work with. Professional, accommodating, organized, knowledgeable ... We could not have been as successful without you.
Senior Manager | Top Four Global Consulting Firm

Partners & Certifications

Ready to unleash the value in your data?