I cancelled Claude. Then I stopped caring about AI quality wars entirely.
Last week, someone published a post called "I Cancelled Claude: Token Issues, Declining Quality, and Poor Support". It hit 392 points on Hacker News. The comments section was a support group.
I've been there.
Not because Claude is uniquely bad — but because every AI subscription eventually disappoints you in the same way.
The pattern always looks like this
- Model launches. Reviews are glowing. You sign up.
- Six months in, the model gets quietly updated. Quality shifts.
- You notice. You open a support ticket. Support is... not helpful.
- Meanwhile, a new model launches. You're told to upgrade.
- Repeat.
Now throw in DeepSeek v4 (1,618 Hacker News points yesterday, 1,270 comments) and GPT-5.5 launching the same week, and you've got a perfect storm of "which AI do I actually trust right now?"
The honest answer? I stopped asking.
What I did instead
I built a thin wrapper around Claude's API with a flat monthly rate. No per-token billing. No tier upgrades. No support tickets about model quality.
Here's the architecture in plain terms:
// The entire business logic is this simple
const response = await anthropic.messages.create({
model: 'claude-3-5-sonnet-20241022',
max_tokens: 8192,
messages: conversation.slice(-10) // keep context window lean
});
// Flat rate covers the API cost. No surprise bills.
// User pays $2/month. I pay ~$0.80/month in API costs.
// The rest funds animal rescue.
The key insight: I'm not in the AI quality business. I'm in the reliability business.
When Anthropic changes Claude's behavior (and they will — they always do), I update the model string. My users don't notice. There's no subscription tier to explain, no quality regression to apologize for.
Why the "cancelled Claude" complaints make sense
The people rage-quitting Claude aren't wrong about the quality issues. They're wrong about what they signed up for.
They signed up for a product relationship with Anthropic. That means:
- Their model roadmap affects your workflows
- Their pricing changes affect your budget
- Their support quality affects your productivity
- Their Claude Code removal from Pro tier affects your tools
When you go direct-to-API with a flat-rate wrapper, that relationship disappears. You're buying electricity, not a relationship with the power company.
The DeepSeek v4 problem
DeepSeek v4 is genuinely impressive. The HN thread is full of real benchmarks and real use cases.
But here's the thing: I'm not going to migrate my users to DeepSeek v4. Not because it's bad — but because the migration cost is non-zero and the improvement is marginal for everyday use cases (summarization, Q&A, writing assistance).
Every "better model launches" moment is a hidden tax on every AI product that made model quality its core value proposition.
Flat-rate wrappers don't have that problem. The model is an implementation detail.
The actual numbers
I run SimplyLouie at $2/month. Here's what that covers:
- ~4,000 messages/month at moderate length
- Claude 3.5 Sonnet (current) as the backbone
- No rate limits that affect normal usage
- 50% of revenue goes to animal rescue
Is it the bleeding edge? No. Is it reliable, predictable, and not going to send you a surprise $80 bill? Yes.
For most use cases — drafting, Q&A, summarization, debugging help — the difference between Claude 3.5 Sonnet and whatever just launched is smaller than the difference between having a tool and rage-quitting because support didn't respond.
What I actually use each model for (right now)
Since everyone's asking:
- Claude 3.5 Sonnet (via SimplyLouie): Daily writing, summarization, Q&A. Reliable, no surprises.
- DeepSeek v4: When I need to benchmark something new. Not production.
- GPT-5.5: When a client specifically requires it. Their bill, not mine.
That's it. I don't track every launch. I don't read every benchmark thread. I made that decision once and stopped revisiting it.
The question I keep coming back to
The HN thread on "I Cancelled Claude" has 208 comments. A lot of them are people describing the same workflow: sign up, get burned, cancel, try something else, repeat.
My question for the community: has anyone actually found a stable AI workflow that survived two model generation changes without requiring a significant migration or repricing?
I'm genuinely curious whether "pick a model and commit" is viable, or whether the churn is structurally inevitable at the product layer.
SimplyLouie is a $2/month AI assistant. 50% of revenue goes to animal rescue. No model-quality treadmill. simplylouie.com
Top comments (0)