Free
Open source, MIT license
Run LLMs locally. Free, open-source. No API bills. Cost is hardware and power.
Free software. No per-token costs. You pay for: GPU/hardware, electricity, and maintenance.
Free
Open source, MIT license
One-off
GPU required; 8GB+ VRAM for smaller models, 24GB+ for 14B, 40GB+ for 70B
Ongoing
See our Self-hosted GPU Comparisons for £/1M token estimates
When it makes sense: Good for data-sensitive workloads, high volume where cloud costs would exceed hardware, or air-gapped environments. Compare with our GPU cost guide.
All data stays on your hardware. No data sent to third parties. Full control.
GDPR / compliance. Data never leaves your infrastructure.
Data sovereignty. Complete data sovereignty.
Every team's fit is different. We'll model cost and ROI across cloud, self-hosted, and hybrid before recommending anything, including this product.