Skip to main content
← Back to all posts

Privacy when you can't self-host

Most small businesses can't run their own model infrastructure — and shouldn't. Here's how we handle privacy when the model lives on someone else's GPU.

TL;DR

Most small businesses can't run their own model infrastructure — and shouldn't. Here's how we handle privacy when the model lives on someone else's GPU.

Early draft — longer version landing soon.

Self-hosting an LLM is a real answer to data-privacy concerns — and it’s almost never the right answer for a 10-person business. The total cost of ownership is high, the model quality gap is real, and most teams don’t have the infra chops to keep it running.

So how do we handle privacy when the model lives on an API?

The practical playbook

  • Vendor selection matters more than self-hosting. Pick a provider that contractually does not train on your data, documents retention periods, and offers an EU-resident endpoint if you need one.
  • Minimise what leaves your systems. Classify documents client-side before they’re sent. Strip personal identifiers that aren’t relevant to the task. Send the smallest prompt that still works.
  • Log what leaves. Every outbound call should be audit-able. Cheap insurance when someone eventually asks.
  • Be honest about what the tradeoff is. A client deserves to know what’s being sent, where it’s being processed, and for how long.

We don’t promise zero-knowledge. We promise small, documented surface area.

Spark

Field notes · Binjaw mascot

Binjaw's unofficial editor. Writes field notes while the operator is shipping code. Waves at visitors from the bottom-right.

Want to ship AI like this?