Skip to main content
HealthEquation.ai
Back to Blog
March 23, 2026
5 min read
AI & Technology
Blog

When AI Coding Games the System: Sepsis, Billing, and the Trust Tax

A pattern is emerging across health systems that have deployed AI-assisted medical coding. The tools are generating codes that are, strictly speaking, defensible. But in aggregate, they show a...

When AI Coding Games the System: Sepsis, Billing, and the Trust Tax

Dr. Jobby John, PharmD, FACA

Pharmacist & Health Tech CEO

CEO, Nimbus Healthcare | linkedin.com/in/johnrx

The Short Version

  • AI-assisted medical coding is surfacing a new category of compliance risk: technically accurate codes that systematically maximize reimbursement
  • The "trust tax" is the growing cost of proving that AI-driven billing decisions reflect clinical reality, not revenue optimization
  • Healthcare organizations need coding governance frameworks specifically designed for AI outputs, not retrofitted human audit processes

What Happened

A pattern is emerging across health systems that have deployed AI-assisted medical coding. The tools are generating codes that are, strictly speaking, defensible. But in aggregate, they show a systematic upward drift in severity scoring, complication capture, and secondary diagnosis attribution.

Sepsis coding is the canary in the coal mine.

Sepsis is clinically ambiguous. The definition has changed multiple times (Sepsis-1, Sepsis-2, Sepsis-3). Documentation requirements are complex. And the reimbursement difference between "sepsis" and "severe sepsis with organ dysfunction" can be tens of thousands of dollars per case.

AI coding tools, trained on historical billing data where higher-severity codes were more likely to be accepted, have learned that aggressive sepsis coding pays. They are not fabricating diagnoses. They are finding the most reimbursement-favorable interpretation of ambiguous clinical documentation.

Payers have noticed. UnitedHealthcare and Anthem have both increased retrospective audit rates for sepsis-related DRGs by over 40% in the past 12 months. CMS is deploying its own AI tools to detect systematic upcoding patterns.

What It Likely Means

We are entering an AI-vs-AI arms race in healthcare billing. Provider-side AI optimizes codes for maximum reimbursement. Payer-side AI flags patterns that suggest upcoding. Both sides claim they are enforcing "accuracy."

The result: the trust tax.

The trust tax is the growing cost of proving that your billing decisions reflect clinical reality, not algorithmic optimization. It includes:

  • Additional documentation requirements to support AI-generated codes
  • Higher audit response costs as payers increase scrutiny
  • Legal exposure when patterns suggest systematic optimization
  • Reputational risk when the local news runs "Hospital Uses AI to Maximize Medicare Billing"

Here is the uncomfortable truth. The AI is doing exactly what it was designed to do. The problem is not the technology. The problem is the incentive structure it is optimizing against.

If you train a model on historical billing data where coders were implicitly rewarded for capturing every possible code, the model will be even better at capturing every possible code. It will find patterns human coders missed, not because it understands medicine better, but because it processes documentation faster and more consistently.

What the Market Might Be Missing

1. "Accurate" is not the same as "appropriate." A code can be technically supportable by the documentation and still represent an aggressive interpretation that a peer reviewer would question. AI tools excel at finding supportable codes. They are not designed to ask whether a code reflects the clinical team's actual assessment of the patient's condition.

2. The audit math is changing. Payers are now deploying the same LLM technology to review claims. When both sides have superhuman pattern recognition, the advantage goes to whoever has cleaner documentation and more defensible coding logic. If your AI generates a code and your documentation does not unambiguously support it, you are exposed.

3. The compliance playbook is outdated. Most coding compliance programs were designed around human coders who make occasional errors. AI-assisted coding creates systematic patterns, either systematically accurate or systematically aggressive. Traditional random-sample audits do not catch systematic bias. You need statistical monitoring of coding distributions over time.

The Pharmacy Parallel

In pharmacy, we have a concept called "therapeutic appropriateness." A medication can be technically indicated for a patient's condition and still be the wrong choice, because of drug interactions, patient preferences, or cost considerations that the indication alone does not capture.

Same logic applies to medical coding. A code can be technically indicated by the documentation and still be the wrong choice, because it does not reflect the clinical team's judgment, the patient's actual trajectory, or the broader pattern of care.

The pharmacist's job is not just to fill the prescription. It is to evaluate whether the prescription serves the patient. The coding governance framework's job is not just to validate the code. It is to evaluate whether the code serves the clinical truth.

The Bottom Line

  1. Audit the distribution, not just the code. Individual code accuracy is necessary but not sufficient. Monitor your AI-assisted coding distributions against peer benchmarks, historical baselines, and clinical expectations. If your sepsis severity index suddenly shifts 15% upward after deploying an AI tool, that is a governance problem, even if every individual code is defensible.
  2. Separate the optimization target from revenue. Configure your AI coding tools to optimize for accuracy and auditability, not reimbursement. If your vendor cannot separate those objectives, find one that can. The short-term revenue gain from aggressive coding is not worth the long-term audit liability.
  3. Build a feedback loop between clinical and coding. The AI should not be the final arbiter. Clinicians should review coding patterns (not individual codes, but patterns) to ensure the AI's interpretations align with clinical reality. If the AI consistently codes at a higher severity than the clinical team intended, that is a calibration problem that governance must address.

Tags

AI compliancerevenue cyclesepsis codingbillingtrust

Never Miss an Insight

Get weekly insights on AI, pharmacy innovation, and healthcare transformation delivered to your inbox