DEV Community

Cover image for My Team Tracks AI-Generated Code. The Number Shocked Us.
Aditya Agarwal
Aditya Agarwal

Posted on

My Team Tracks AI-Generated Code. The Number Shocked Us.

My team tracks how much of our codebase is AI-generated. The number shocked us.

We deployed Buildermark last week. It's an open-source tool that scans Git history and flags AI-written lines.


Why We Started Measuring

Every startup has that moment.

You're reviewing a PR and realize you can't tell who wrote it. The human or the AI.

We hit 40% AI-generated code by volume. Some files were 90%.

The CTO asked for the report. Then asked what it meant.

Nobody had an answer.


The Three Problems Nobody Talks About

Problem 1: Ownership blur

When AI writes the fix, who owns the bug?

We found junior devs treating Claude output as gospel. They'd copy-paste without understanding.

Senior engineers would approve because "it looks fine."

Problem 2: The review gap

Human-written code gets scrutinized. AI-written code gets rubber-stamped.

We caught security issues in AI-generated config files. Stuff a human would never write.

Problem 3: The bus factor

If your AI provider degrades (like Claude did last month), your velocity tanks overnight.

We're now vendor-locked to Codeium's style. Claude's patterns. GitHub Copilot's idioms.


What We Changed This Week

We added a pre‑commit hook that tags AI‑generated lines.

Every PR shows the percentage in the description.

If it's over 50%, it needs extra review. No shortcuts.

We also started tracking "AI debt" – lines that only one person understands because they came from a prompt nobody wrote down.


The Real Metric That Matters

Lines of AI code is vanity.

The real metric is: How many AI‑generated lines survive to production without a human understanding them?

We're at 12%.

That's 12% of our codebase that could break and nobody would know why.


Is your team measuring AI code?

What percentage would surprise you?

👇

Top comments (0)