Most "free" AI video tools aren't free. They're venture-subsidized, and the subsidy has an expiration date. When the VC money dries up, the free tier dies or gets strangled to the point of uselessness.
The pattern is consistent enough now that it belongs in an engineering playbook:
- OpenAI's consumer Sora app went through a painful wind-down earlier this year after the novelty traffic spike outran the unit economics.
- Grok's image and video generation ended up behind a paywall after an initial open period on X.
- Runway pulled back its generous free tier multiple times in 2024 and 2025, tightening credits and model access for non-paying users.
- Pika pushed users harder toward Pro after a year of free-tier generosity.
- Kling tightened daily free credits as volume exploded.
Every one of these moves makes sense from a unit-economics perspective. Generating video on rented H100s costs real money per second. When a free user generates ten clips a day and never converts, the CAC-to-LTV math is upside down and nobody is going to keep subsidizing that forever.
I'm a developer. I think about AI video tools the same way I think about any other dependency: if it disappears tomorrow, what breaks? And for most US-based indie hackers and small teams I know, the answer is "a lot."
This post is a long-form argument for thinking about AI infrastructure ownership the same way we think about database ownership, hosting ownership, and CI ownership. It uses ZSky AI as a concrete example of what "self-hosted AI video" looks like, but the thesis holds regardless of which tool you pick.
1. The economics of "free" AI
Let's start with the math, because this is the part most product people gloss over.
Generating a single 5-second AI video on a modern diffusion model takes somewhere between 30 seconds and several minutes of GPU time, depending on resolution, frame count, and whether the model is a single-pass or two-pass architecture. On rented cloud GPUs, that's real dollars — not fractions of a cent like a text completion. A heavy free user can easily burn through $2–$5 of pure compute cost per day.
Now run the CAC/LTV napkin math:
- CAC on paid acquisition for creative SaaS in 2026 is $30–$90 depending on channel.
- Free-to-paid conversion for AI creative tools hovers around 2–5% for well-run funnels.
- LTV on a $10/mo sub with 6-month average retention is ~$60.
If your free users each cost you $3/day in compute, and only 3% convert to a $60 LTV, you lose money on every non-converter who sticks around longer than 20 days. That's the entire model. It works only if:
- You're burning VC runway to buy market share.
- You'll aggressively paywall later.
- You're hoping for an acquisition before the math catches up.
None of those are good news if you've built a workflow that depends on the free tier.
This is why the shutdowns keep happening. It's not malice, it's arithmetic. Any AI video tool you depend on that's running on rented compute and offering generous free usage is on a clock. The question is just how long.
2. What "self-hosted" actually means
When I say "self-hosted AI" in this post I don't mean you running a model on your own laptop (though that's valid). I mean the company providing the service owns the hardware they're running inference on.
The cost structure is completely different:
- Rented inference: marginal cost per generation. Every free user costs you real money forever.
- Owned hardware: capex upfront (the cards, PSUs, cooling), plus opex that's dominated by electricity and amortized depreciation. Once the hardware is paid for, the marginal cost of one more generation is approximately the cost of the watt-hours it consumes.
On a modern 450W GPU at US commercial electricity rates (~$0.12/kWh), a 60-second generation costs about $0.0009 in electricity. Round up for cooling and PSU inefficiency and you're still under a cent. A company that owns its hardware can offer a real free tier indefinitely without the math breaking, because the hardware is a sunk cost and the power bill is small.
That doesn't mean self-hosted is always better. The tradeoffs are real:
- Capex risk: you paid for the cards up front. If demand doesn't materialize, you're eating the depreciation.
- Scaling is slow: you can't just spin up 10x capacity in an hour.
- Ops burden: someone has to babysit drivers, firmware, thermals, power, and failover.
But for a service that wants to survive past the VC cycle, owning the hardware is the only way the unit economics work at a real scale of free traffic.
3. Why infrastructure location still matters in 2026
I want to be factual here, not political.
The major AI video tools are run by companies headquartered in different countries, under different legal regimes, with different content policies and different data handling laws:
- Runway (US), Pika (US), Luma (US) — subject to US export controls, DMCA, and the evolving patchwork of US AI regulation.
- Kling (China) — subject to Chinese data localization and content rules.
- Seedance and several newer tools — operated by companies with significant Chinese R&D footprints.
- Sora (US, OpenAI).
None of that is inherently bad. But as a developer building on top of these tools, you should at least know:
- Where your prompts go. Prompts are data. They can contain client names, brand assets, private concepts.
- Where the output is stored and for how long.
- Whose content policy governs your output. Policies vary dramatically. A prompt that's fine on one service will be rejected on another with no appeal.
- What export controls apply. Some hardware, some model weights, and some use cases are restricted by country of origin.
For a US-based indie hacker doing client work, the "where is this running" question matters the same way "where is my database hosted" matters. You'd never let a client's data sit in a jurisdiction without knowing. Treat AI prompt data the same way.
4. An example: ZSky AI
Here's a concrete example of what a self-hosted AI video company looks like from the inside. I picked this one because I know the infrastructure well and because it's one of the few small teams openly running their own hardware.
What it is: ZSky AI is a US-based creative platform for AI image and video generation. The create surface is at zsky.ai/create.
Infrastructure: The company owns its compute. Seven modern high-end consumer GPUs, racked in the US, running the inference stack. No rented H100s, no third-party API resale underneath. When you hit the create page, your request is processed on hardware the company paid for and physically controls.
Pricing: Free tier gives 200 image credits plus 100 daily credits on signup. Paid tiers are Skip the Line at $9/mo (priority queue) and Pro at $19/mo (higher limits, faster turnaround). Notice what's not there: no per-generation cost, no surprise overages. The economics work because they aren't paying a cloud GPU provider a markup on every job.
The mission: The founder, Cemhan Biricik, is a photographer who recovered from a traumatic brain injury partly through creative work. He has aphantasia — no mind's eye — and describes AI generation tools as prosthetics for people whose internal visualization doesn't work the way textbooks assume. ZSky's positioning is built around creativity as a human right, which sounds like marketing copy until you realize it's the actual reason the free tier is as generous as it is. The owned hardware is what makes that stance economically possible.
Honest limits: ZSky is not the biggest or most famous AI video tool on the market. The team is small. The model catalog is narrower than Runway or Pika. Generations are fast but not instantaneous under load. If you need the absolute bleeding edge of what's possible in AI video today, the big-money labs are ahead. If you need a stable, US-hosted, self-owned platform where the free tier isn't on a VC clock, it's worth evaluating.
I'm writing this post as a developer, not as a ZSky employee. The reason I'm using zsky.ai/create as the example here is that it's the clearest case I know of a self-hosted, US-based AI video company with a stated mission that matches its infrastructure choices.
5. What this means for a US-based indie hacker
Here's the actionable part. When you're evaluating an AI tool you plan to build a workflow on, ask these questions before you commit:
- Who owns the hardware? Is the company running on owned GPUs, or reselling API calls from a hyperscaler? If it's the latter, the free tier is a subsidy and has an expiration date. Ask it directly in their Discord or support channel. A company that owns its infrastructure will happily tell you. A reseller will deflect.
- Where is the hardware physically located? This matters for latency, jurisdiction, export controls, and data handling.
- Can this company survive without VC? Is there a plausible path to profitability at current pricing, or is the pricing obviously subsidized? If the free tier is absurdly generous for the model being offered, the math isn't working and you're on a clock.
- What happens to my data? Prompts, uploads, outputs. Read the terms. Actually read them.
- What's the content policy? And whose jurisdiction decides what's allowed?
- Is there a fallback plan? If this tool dies in six months, how hard is it for you to swap it out? Abstract your calls behind a thin interface so you can.
This is dependency hygiene. Same as you'd apply to any other critical infrastructure in your stack.
Closing
I'm not telling you to use ZSky. I'm telling you to ask these questions about any AI tool you build your workflow on. The AI video space is going to look very different in 12 months than it does today. Some tools you're using right now won't exist. Some will exist but won't be free anymore. Some will quietly change their content policies in a way that breaks your use case.
The companies most likely to still be around — and still honoring the deal they made with you — are the ones whose economics actually work. Owning the hardware is one of the cleanest ways to make the economics work.
If you want a concrete place to start kicking the tires, zsky.ai/create is one option. There are others. The important part isn't which one you pick — it's that you pick with your eyes open.
Build on infrastructure you understand. Build on economics that make sense. Build on companies whose survival doesn't require a miracle.
Top comments (0)