← Back to Blog
Healthcare AI · Compliance

Building HIPAA-Compliant AI: What Developers Get Wrong

Claudeter Team February 3, 2026 10 min read
Share

Every engineering team building AI for healthcare will tell you their system is HIPAA compliant. Ask them how, and most will describe the same three things: data encryption at rest and in transit, access controls with authentication, and a BAA signed with their cloud provider.

These are necessary but not sufficient. HIPAA compliance in AI systems is harder than compliance in traditional software — and most engineering teams are failing at the parts that aren't obvious.

The Obvious Parts (That Everyone Gets Right)

Good. Check these boxes. But these are baseline — table stakes for any software handling PHI, AI or not.

The Hard Parts (Where Most Teams Fall Short)

Model Inference and PHI

When you send PHI to an LLM for processing, where does that data go? OpenAI's enterprise tier has BAAs and specific data handling commitments. Their consumer API does not. Anthropic's enterprise offering has specific HIPAA provisions. Their standard API tier is different.

The failure mode: an engineering team signs a BAA with their cloud provider (AWS), hosts their application on AWS, but sends PHI to the OpenAI API on the standard tier. The cloud hosting is compliant. The inference is not. The BAA doesn't cover the LLM call.

Every external API call that touches PHI needs its own BAA. Model inference is an API call. This is the compliance gap most AI healthcare teams have.

Audit Logging at the Right Granularity

HIPAA requires audit controls — a log of who accessed what PHI and when. Most teams implement application-level logging. What they miss is logging at the model inference level: what PHI was included in which prompt, what the model returned, and what action was taken based on that response.

This matters for breach assessment. If your model inference logging is insufficient, you can't reconstruct exactly what PHI was exposed in a breach scenario — which complicates both the breach notification analysis and any subsequent regulatory review.

De-identification vs Pseudonymization

There's a meaningful legal difference between de-identified data (which is no longer PHI under HIPAA's Safe Harbor or Expert Determination methods) and pseudonymized data (which still is PHI because it can be re-identified with a key). Many AI teams treat these as equivalent when fine-tuning models on healthcare data. They're not.

3
obvious compliance areas
7+
gaps most teams miss
$50K+
min HIPAA violation fine

Data Residency for UAE and Global Deployments

For UAE deployments, the DHA has specific data residency requirements — patient data must remain within UAE borders. This creates a challenge for AI systems that rely on US-based model inference APIs. Solutions include UAE-hosted model deployments (Azure UAE North, AWS ME-South-1) or on-premises inference for sensitive operations.

Building It Right From the Start

The most expensive HIPAA compliance fix is a retrofit. A system built without proper audit logging takes weeks to add it correctly. A system that has been sending PHI to non-BAA-covered APIs has a retroactive exposure problem that's difficult to resolve cleanly.

The right time to address compliance architecture is in week 1 of development — not in the week before launch when a compliance review catches the gaps. This is why we treat HIPAA compliance as a feature requirement, not an audit checklist.

The Minimum Viable Compliance Stack

This isn't a comprehensive HIPAA compliance program — that requires a full risk assessment and security officer involvement. But it's the technical baseline that most AI healthcare teams are missing.

Ready to Automate?

Talk to our team about building a custom AI solution for your workflow. POC in 3 days, live in 6 weeks.

Book a Free Discovery Call →