When Real-World Data Breaks Software Assumptions
- Sherwin Gaddis

- Feb 26
- 2 min read
Summary Video:
Most EHRs work well in theory.
They pass demos.
They pass checklists.
They pass feature comparisons.
Then real data shows up.
That’s usually when the assumptions break.
The quiet assumption most systems make
Most EHRs are built around an unstated belief:
Data will arrive in reasonable volumes, in predictable shapes, and be accessed occasionally.
That assumption holds early on.
It rarely survives daily use.
Labs are a good example.
On paper, lab integration sounds simple:
place an order
receive a result
display it in the chart
In practice, labs don’t return a result.
They return streams of data — large, continuous, and unevenly relevant over time.
Systems designed for forms suddenly have to behave like data managers.
What actually happens during onboarding
During a recent onboarding, lab ordering was configured just enough to send orders and receive results.
Technically, everything worked.
Operationally, it didn’t.
Once real lab data began flowing:
Volumes were higher than expected
Result sets were broader
Relevance depended on context
Providers needed control, not just access
The system wasn’t broken.
It was behaving exactly as designed.
The problem was that the design assumptions no longer matched reality.

Why “just show the data” fails
A common reaction at this stage is:
“Can we just show all of it?”
That approach rarely scales.
Unfiltered data creates:
cognitive overload
slower documentation
missed signals
frustration that gets labeled as “performance issues.”
But the issue isn’t speed.
Its structure.
Real-world clinical systems require:
controlled access
meaningful grouping
context-aware display
intentional friction in the right places
Without structure, more data makes work harder, not easier.
The work that starts after “it works.”
This is the phase most systems are never designed for.
Once data flows correctly, the real questions emerge:
Who needs to see this?
When do they need to see it?
What should remain hidden until requested?
What repeats often enough to deserve structure?
These aren’t feature questions.
They’re operational ones.
They only surface when a system is in active use.
Why this matters more than features
Most EHR comparisons focus on capability:
Does it integrate?
Does it support the order?
Does it store the result?
Those questions are baseline requirements.
What determines whether a system holds up over time is:
how it behaves under real data
how quickly assumptions can be adjusted
how much friction can be removed without breaking compliance
That’s not about adding features.
It’s about reshaping the system to reflect how work actually happens.

A pattern we’ve learned to expect
This isn’t an edge case.
It’s a pattern.
A system goes live
Real data arrives
Assumptions are exposed
Structures are adjusted
Work becomes smoother
The clinics that struggle aren’t using “bad software.”
They’re using systems that can’t adapt once reality shows up.
Closing thought
Software rarely fails on day one.
It fails later — quietly — when volume, pace, and complexity arrive.
The difference between a system that frustrates and one that endures isn’t polish.
It’s whether someone is watching closely enough to notice when reality no longer matches the assumptions — and is willing to rebuild from there.


Comments