Back to all essays
Own Your Tech

I Self-Hosted My Own AI for $700 a Month. Here's What I Actually Learned

·9 min read
George Pu
George Pu$10M+ Portfolio

27 · Toronto · Building businesses to own for 30+ years

I Self-Hosted My Own AI for $700 a Month. Here's What I Actually Learned

TL;DR:

  • Full self-hosting of top-tier AI costs $20,000–$50,000/month and isn't worth it for most businesses.
  • Cloud-hosted models (AWS Bedrock, Google Vertex, Azure) are the underrated middle ground — you get the best AI with your data staying in your account.
  • I built a private AI setup for $700/month and it covers 80% of daily tasks. Here's the full breakdown, the real costs, and a decision framework for what makes sense for your business.

My team and I just spent a few days building our own AI system from scratch. Hosted it on our own cloud server. Ran it alongside the AI tools we use every day — Claude, ChatGPT, all of them.

The goal wasn't to replace those tools. It was to answer a question that's been nagging me: what does it actually take to own your AI and your data?

Short answer: it's more accessible than you think, less powerful than you'd hope, and the real play is somewhere in the middle that almost nobody is talking about.

Let me walk you through everything we found.

The Experiment: What We Actually Built

We rented a server from Google Cloud in the Toronto region (Canada — this matters for where your data physically lives), equipped with a specialized AI chip.

On that server, we installed an open-source AI model called Qwen3 — think of it as a free, publicly available alternative to ChatGPT — and put a familiar chat interface on top so anyone on the team could use it like they'd use ChatGPT or Claude.

Total setup time: a few hours. Total monthly cost: about $700.

And honestly? I was shocked. It just worked.

Not "worked" as in it competes with Claude or GPT frontier models. It doesn't. This is a smaller, less capable model. On complex reasoning, deep analysis, nuanced writing — the top-tier tools are clearly better.

But here's the thing. The tasks that make up 80% of daily AI usage — drafting emails, summarizing documents, sorting through data, answering internal questions, basic code generation — it handles them fine. Responses come back quickly. Quality is good enough.

I think most of us who work with AI daily have gotten spoiled. We use the most powerful models on the market for everything because the subscription is flat-rate. There's no extra cost to asking the most advanced AI ever built to sort your inbox. But you don't need it for that. You're driving a Ferrari to the grocery store.

Even Anthropic — the company behind Claude — figured this out. Their coding tool uses their smallest, cheapest model as a helper for routine work. If you've seen how fast and efficient that is, you already know: smaller models are shockingly capable for everyday tasks.

We take these frontier models for granted because we're paying flat-rate. But when you actually break it down by task, most of what we do doesn't require the best model on earth.

Why I Even Started Thinking About This

I subscribe to every major AI platform. Claude, ChatGPT, Perplexity, Gemini, Grok — if it exists, I'm probably paying for it.

And I've gotten comfortable giving my data to all of them. Maybe too comfortable.

Here's what sits in the back of my mind: with social media, we gave our data to Google, Facebook, TikTok. We accepted it because what we were sharing — photos, status updates, likes — felt low-stakes.

With AI, it's different. Way different.

We're sharing company strategies. Financial models. Hiring plans. Investor conversations. Legal questions. Personal health concerns. The intimacy of what goes into these chat windows is on a completely different level than anything we've shared with tech platforms before. And we're doing it casually, without thinking twice.

Yes, I opt out of training wherever I can. But let me be real — the risks are there. These companies retain your data. Moderators may be flagging and reading conversations that trigger certain keywords. Humans might be reviewing your chats without you knowing. All perfectly legal under the terms and conditions you clicked "accept" on without reading.

Even on enterprise plans — the expensive corporate versions — your data still sits on their servers. More guardrails, sure. But the data doesn't vanish. It's still there, on someone else's infrastructure, under someone else's control.

So I wanted to know: what does it actually cost to keep it all under your own roof?

The Brutal Truth About Full Self-Hosting

Let me be real — if you want to host an AI that actually competes with the best commercial tools, the math is ugly.

Here's why: the AI models you use through a $20 or $200/month subscription are running on massive clusters of specialized chips that cost tens of thousands of dollars each. You're sharing those chips with millions of other users. The economics work because the cost is spread across everyone.

When you self-host, you're renting or buying those chips just for you. And the bill reflects that.

The closest open-source model to the top commercial ones right now needs 8 to 16 of the most powerful AI chips available. Running that 24/7:

Minimal setup: $15,000–$25,000 per month with discounted pricing.

Full production setup: $25,000–$50,000 per month depending on provider and configuration.

I'll admit — I underestimated these numbers before we started researching. I knew it was expensive. I didn't know it was that expensive.

And that's just for the system to be running. There's another problem most people don't think about at all.

The Speed Problem Nobody Mentions

This one caught me off guard too.

When one person is using a self-hosted AI, it's fast. Responses stream back in seconds.

When 15 people use it at the same time? Still usable. A bit slower, but fine.

50 people simultaneously? You start feeling it. Responses take noticeably longer.

100 people at once? Everyone's waiting. The system is crunching a lot in total, but each individual person's experience gets worse. Noticeably worse.

So if you're a company of 500 employees and 50–100 of them are actively using AI at any given time, you need multiple clusters of high-end chips to keep things feeling responsive. That pushes costs to $50,000–$120,000 per month.

Now compare that to enterprise AI plans from the major providers. Based on reported pricing, OpenAI's enterprise plan runs around $60 per user per month. Anthropic's Claude enterprise is similar. For a 100-person company, that's roughly $72,000–$100,000 per year. Not per month. Per year.

On price alone, self-hosting is dramatically more expensive than just paying for enterprise subscriptions.

What you gain is data sovereignty. Your conversations, documents, strategies — they never leave your infrastructure. For most companies, that trade-off doesn't make financial sense. For some, it's non-negotiable. Know which one you are.

The Middle Ground Nobody Talks About

Okay, this is the part I think matters most. And the part almost every guide on this topic skips entirely.

There's a whole tier between "send everything to OpenAI's servers and hope for the best" and "spend $50K a month building your own AI fortress." It's called cloud-provider hosted models. And for most businesses, this is probably the right answer.

Here's the concept in plain terms:

Want the full playbook? I wrote a free 350+ page book on building without VC.
Read the free book·Online, free

The Big Cloud Providers Now Host AI Models For You

Amazon, Google, and Microsoft have each partnered with the major AI companies. Instead of your data going directly to OpenAI or Anthropic, the AI model runs inside your cloud provider's infrastructure — in the region you choose, under your account's security policies.

Amazon Web Services + Anthropic (Claude):

When you use Claude through AWS, the model runs inside Amazon's data centers, not Anthropic's. You pick the region — I'm in Toronto, so I select Canada. My data stays in Canada, inside my AWS account. Amazon states that customer content is not used to train the AI models. Your stuff doesn't touch Anthropic's servers.

Google Cloud + Various AI Models:

Same idea. You deploy AI inside your Google Cloud account, in your chosen region. You control access. Your data stays in your project unless you explicitly send it elsewhere.

Microsoft Azure + OpenAI:

Same arrangement. You access OpenAI's models through Microsoft's infrastructure without data going to OpenAI directly. If something goes wrong, Microsoft is your provider with enterprise contracts and accountability. You have someone to call.

What This Actually Gives You

Your data stays in your cloud account, in your chosen geographic region, under your security policies and encryption. You get the full power of the best AI models available, without your sensitive information flowing through the AI company's own servers.

Is it the same as owning the physical hardware yourself? No. The cloud provider still operates the platform. And yes — there's a theoretical risk. A US government data request could potentially reach AWS or Google infrastructure even if it's hosted in a Canadian region. That's real, and worth knowing about.

But for 99% of business use cases, this is more than enough. You get top-tier AI quality, your data stays where you want it, and you're not hemorrhaging $50,000 a month on specialized chips.

This is the option I think most businesses should seriously look at. If you're a government entity or defense contractor with strict sovereignty requirements, you might need full self-hosting. For almost everyone else, this middle ground is your answer.

I've worked directly with both AWS and Google Cloud on setups like this. The experience is good. This isn't experimental — it's production-ready infrastructure that major enterprises are already using today.

What Our $700/Month Setup Actually Looks Like

Back to our experiment. We weren't trying to replace the best AI tools. We wanted a private AI layer — something for sensitive work that we don't want flowing through any third-party, period.

Here's the setup:

  • Where it runs: Google Cloud, Toronto (Canada) region
  • Hardware: A single server with a mid-tier AI chip
  • AI model: Qwen3 — an open-source model (for context, the top commercial models are roughly 10-50x larger)
  • Interface: An open-source chat interface that looks and feels like ChatGPT
  • Monthly cost: ~$700

For a small team of 3–5 people, it handles daily workloads without issues. The responses aren't as sophisticated as Claude or ChatGPT at their best, but for most routine work? They don't need to be.

We use this for things where we want the data staying on our servers. We use frontier models for everything else. Best of both worlds.

That's the key insight here — you don't have to pick one. Run both. Use the right tool for the task. Stop using the Ferrari for groceries.

The Even Cheaper Path: Just Buy the Hardware

If you want to go even further, skip the cloud entirely.

Apple's Mac Mini Pro ($1,500–$2,000) can run capable open-source AI models locally, 24/7, right from your desk or office. Responses will be slower — maybe 10–30 seconds for complex questions — but it works. After the purchase, you're paying a few dollars a month in electricity. That's it.

You can also build a dedicated AI server with a specialized chip for about $3,000–$4,000. That gives you a permanent, always-on AI system that you own completely. No monthly bills. No cloud provider. Just yours.

Compare that to $700/month in cloud costs: you break even in 5–6 months. After that, it's essentially free. You own the hardware, you own the data, nobody gets a monthly check from you.

We're planning to try this next — building a physical AI server in-house. It's part of what we're exploring with Sovereign Cloud. If that interests you, I'll share what we learn.

So What Should You Actually Do?

Here's my honest framework. No fluff.

If you're a small team (under 10 people) doing general work:

Just use Claude or ChatGPT. The $20–$200/month subscriptions are the best deal in tech right now. You're accessing systems that cost billions to build. Don't overthink this.

Look at the cloud-hosted options — AWS Bedrock, Google Vertex, Azure OpenAI. Best AI available, your data stays in your account. This is the sweet spot for professional services firms, consultancies, financial institutions, anyone in a regulated industry. Seriously — look at this. It's what I'd do if I were running a 50-person firm.

If you want a private AI alongside your existing subscriptions:

Do what we did. Spin up a small open-source model on a cloud server or buy the hardware outright. $700/month or a one-time $3,000–$4,000 investment. It handles 80% of routine tasks. Use the top-tier models for the other 20%. You'll be surprised how well this works.

If you're a government entity or in a strict regulatory environment:

Full self-hosting on your own infrastructure. Expensive ($20,000–$50,000+ per month), but for some organizations, there is literally no other option.

If you're a company with 50–500+ employees:

Do the math. Enterprise plans from Anthropic and OpenAI are significantly cheaper per person than self-hosting. Unless you have specific sovereignty requirements, go enterprise subscriptions plus cloud-hosted deployment for sensitive workloads. That combination covers almost everything.

Why This Matters Right Now

Look — AI is becoming the operating system for how companies run. Strategy, communications, code, decisions — more and more of it is flowing through AI. And right now, most of that is flowing through a handful of providers who can see everything.

I'm not saying that's evil. These are good companies building incredible products. I pay for them. I'll keep paying for them.

But the question of who controls your AI infrastructure is going to matter. A lot. It's the same shift we saw with cloud computing a decade ago. The companies that understood it early and moved strategically had an advantage. The ones that waited scrambled to catch up when they had no choice.

Same thing is happening now with AI. Except the stakes are higher, because the data is more sensitive.

Figure this out now. Don't wait until the first breach or regulation forces your hand. By then, you're already behind.

Own your AI, or be owned by whoever does.