DEV Community

Akash Melavanki
Akash Melavanki

Posted on

Vercel got hacked. Your API keys rotated. You're still not safe.

I host Thskyshield on Vercel.

So when I woke up to the news that Vercel had been breached — internal systems compromised, customer environment variables exposed, data allegedly being sold on BreachForums for $2 million — I didn't panic. I rotated my keys immediately, like everyone else.

And then I sat with a question that I don't think enough developers are asking:

What happens in the window between a key being stolen and you rotating it?

That question is literally why I built what I built.


What actually happened at Vercel

Let me explain the attack chain, because it's more interesting than "Vercel got hacked."

A Context.ai employee got hit with Lumma Stealer malware in February 2026. Lumma is an infostealer — it quietly harvests credentials, OAuth tokens, session cookies. The attacker sat on those credentials for weeks.

Then they used the stolen OAuth token to access Vercel's Google Workspace. That gave them access to Vercel's internal environments. And in those environments sat thousands of customer environment variables — API keys, database credentials, signing tokens — that weren't marked as "sensitive."

One employee. One third-party tool. One overly permissive OAuth grant. And suddenly an attacker has the keys to a significant chunk of the web's developer infrastructure.

This is what a supply chain attack looks like in 2026. It's not brute force. It's patient, precise, and automated.


The standard advice is incomplete

"Rotate your keys immediately."

Yes. Obviously. Do that.

But here's what nobody's talking about: key rotation is reactive. You rotate after you know you're compromised. The attacker who stole your key at 2 AM on a Sunday doesn't wait for Monday morning. They act the moment they have it.

With AI-powered automation, that window between theft and damage is now measured in seconds, not hours. Google's Threat Intelligence team found that the time between initial access and full breach has collapsed from 8 hours in 2022 to 22 seconds in 2025. The attacker doesn't need you to be asleep. They're done before you finish reading the breach notification email.

So yes, rotate your keys. But also ask: if this key gets stolen tonight, what is the maximum damage the attacker can do with it?


What a stolen LLM API key actually enables

Here's where I want to be honest about scope, because I think most people only think about one dimension of this.

A stolen OpenAI, Anthropic, or Gemini API key gives an attacker several options:

1. Denial of Wallet — loop your chatbot endpoint with high-token payloads. Max out your billing. Leave you with a $5,000 invoice by sunrise.

2. Data exfiltration — if your LLM calls include user data, system prompts, or sensitive context, the attacker can extract that by replaying your own endpoints.

3. Content generation at your cost — use your key as a free compute resource. Generate content, run agents, build products — all billed to you.

4. Prompt injection into your users — if the attacker can make calls to your endpoint, they may be able to manipulate the responses your actual users see.

I want to be clear: I'm only solving one of these. Thskyshield is a financial kill-switch. It stops the billing damage. It does not stop data exfiltration. It doesn't block malicious prompt injection. There are other tools for those layers.

What I believe is: every attack that costs money gets stopped at a known ceiling. If your OpenAI key is stolen and the attacker tries a Denial of Wallet attack, they can only drain up to whatever daily limit you set per user. Not $5,000. Not $500. Whatever you decided.

That ceiling exists even if everything else failed. Even if Vercel leaked your key. Even if the attacker has full access. The financial blast radius is bounded.


Why AI makes this more urgent, not less

I've had this instinct for a while, and the data is starting to confirm it.

Attacks are getting faster because the automation is getting smarter. The Lumma Stealer that hit the Context.ai employee — that's not a human manually harvesting credentials. That's a piece of software running autonomously, finding credentials, exfiltrating them, and handing them off to an operator in real time.

The same automation that makes AI products useful makes AI-powered attacks cheap to run at scale. A Denial of Wallet attack used to require someone sitting at a keyboard, writing a loop script, running it manually. Today it's a five-line agentic task: "Loop this endpoint until the budget hits $X or you get a 429."

The attack surface for every AI product is a financial endpoint. Every chatbot, every AI feature, every LLM-powered tool has a cost function. And attackers are starting to understand that better than most developers do.


The thing I keep coming back to

The Vercel breach is not an outlier. LiteLLM in March. Axios in March. Context.ai in February. Vercel in April.

Every one of these is a supply chain attack. Every one of them exposed developer credentials. Every one of them happened through trusted infrastructure — things developers rely on every day without thinking twice.

You can't stop the breaches. You can't guarantee your keys won't be stolen. What you can control is what happens after.

Rotate your keys — yes. But also put a ceiling on what a stolen key can do to your business.

That's the layer most developers don't have yet.


If you're building with LLM APIs and you want a hard limit on what a stolen key can drain: thskyshield.com

Or if you want to watch a simulated Denial of Wallet attack fire in real time: thskyshield.com/simulator

Curious what others are doing about this. Are you relying on provider-side limits? Rolling your own governance? Or just hoping it doesn't happen to you?

Top comments (0)