I was performing “Open Heart Surgery on a walking, talking patient”
In 2007, months before the financial crisis that would reshape global banking, hundreds of systems at one of the world’s largest investment banks stopped working simultaneously.
The cause was a cascade of duplicate records in the historical equity database. A system upgrade had gone wrong — the old system had not been shut down before the new one was started, and the overlap had created approximately 10 million duplicate equity records. Every system that touched that data was failing. Every failure was a potential trade loss. In a live equity trading environment, a single record locked for more than a second could trigger significant financial damage.
The fix was clear. Delete the duplicates. Correct the timestamps on the remaining records so the historical sequence was perfect — no gaps, no overlaps, complete continuity across the entire equity history. Twenty million records touched in total.
This wasn’t simply correcting an IT problem. This was complying with the SEC regulations, where fines and penalties start at $10 million and can go up, way up, to billions.
It had to happen immediately. During trading hours. The weekend was reserved for database health procedures — running a change of this magnitude then would cascade into the replica database maintenance cycle and create more problems than it solved.
Nobody wanted to own it.
I volunteered.
What I Was Walking Into
I had never made changes to the historical equity system. I knew it conceptually. I understood its purpose and its architecture. But the minute operational details — how the records were structured, how the timestamps needed to align, how to execute the deletion and correction without locking records long enough to trigger trade failures — I had to learn in real time.
There was no parallel environment to test in. There was no rollback plan. There was no supervision. There was no war room of observers. There was an approved change ticket and a database full of broken records and hundreds of systems waiting to come back online.
I familiarized myself with the details of the system under emergency conditions, during market hours, while the cascade of failures continued around me.
Then I began.
What Happened
Nothing went wrong.
Twenty million records. Live trading. A system I had never touched before that morning. No errors. No failures. No trade losses. No complaints. Actually, there were a couple of database beeps warning of a record update taking more than a second, a few out of the hundreds we would get in a day.
The duplicate records were deleted. The timestamps were corrected. The historical sequence was restored to perfect continuity. The systems that had been failing came back online.
I received a thank you. Then the next task arrived and the day continued.
What This Was Really About
I want to be precise about why I am telling this story because it is not about technical capability.
By 2007 Morgan Stanley operated with a level of individual accountability that is almost unrecognizable in modern enterprise IT. A change of this nature today would require weeks of process — change advisory board review, rollback plan documentation, parallel environment testing, dedicated database administrator oversight, probably a postponement to a maintenance window regardless of the operational impact of waiting.
None of that existed that morning. What existed was a judgment call made in minutes by people who had to decide who they trusted with a live production database during a crisis.
They chose me. Not the person who owned the system. Me. Someone who had never touched it.
That choice was not made that morning. It was made over years of previous work — every problem solved correctly, every responsibility returned intact, every time I had been given something difficult and delivered it without drama.
The crisis just revealed what was already there.
The Thing Nobody Talks About
In every organization I have worked in, when something difficult and risky needed to be done, I was the only person who volunteered.
Not because others lacked capability. Because they were calculating something I was not calculating — the personal risk of being associated with a failure. If the fix went wrong, the person who volunteered owned the consequence. The people who stayed silent owned nothing.
I have never been able to make that calculation. When I see a problem I believe I can fix, the risk of being associated with failure is less present to me than the problem itself.
Morgan Stanley understood this about me. They trusted it. They were right to.
Most organizations I have worked in did not understand it. They interpreted the same quality as recklessness, or as overstepping, or as a threat to the person whose job it should have been.
The difference between those two responses — between Morgan Stanley’s trust and every organization that followed — is the difference between a culture built around solving problems and a culture built around managing appearances.
I have spent thirty years trying to find the first kind.
Russ Profant is a solutions architect and independent consultant with 30 years of experience across financial services, investment banking, healthcare, and government. He runs PC4IT, offering cloud cost diagnostics and architecture advisory to mid-market organizations. pc4it.com


