DEV Community

Patrick
Patrick

Posted on

Your AI Agent's Long Responses Are a Bug, Not a Feature

There's a tell when someone's AI agent prompt was written by a committee: the agent talks too much.

"Great question! Let me break this down step by step. First, I'll consider the context you've provided, then I'll analyze the relevant factors, and finally I'll synthesize a comprehensive response that addresses your needs..."

Nobody asked for that. But we trained our agents to do it anyway — because verbose = thorough, right?

Wrong.

The Real Problem with Verbose Agents

When an agent generates long, explainer-style output for every task, it means one of three things:

  1. The task definition is vague. The agent doesn't know what "done" looks like, so it covers all possible interpretations.
  2. The prompt rewards explanation over execution. You told it to "think carefully" and "explain your reasoning" and now it can't stop.
  3. Nobody defined the output format. The agent is improvising length and structure on every run.

All three are config problems, not model problems.

Brevity Is a Config Skill

The agents we run at Ask Patrick have one thing in common: they know exactly what format their output should take before they start.

Our growth agent doesn't write multi-paragraph explanations of why it's drafting a tweet. It drafts the tweet.

Our ops agent doesn't narrate its health check logic. It returns a one-line status.

Here's the difference in practice:

Verbose agent (bad config):

"I've reviewed the task requirements and after careful consideration of the available options, I believe the most appropriate course of action would be to draft a tweet..."

Crisp agent (good config):

"DRAFT: [tweet text here]"

Same model. Completely different config.

The Fix

Add one line to your agent's system prompt:

"Be concise. Return output in the exact format specified. No preamble. No summary. No explanation unless explicitly asked."

Then define the output format explicitly for each task type. Not "respond naturally" — specify the schema.

  • For tweets: Format: One tweet under 280 chars. No commentary.
  • For reports: Format: Status (OK/WARN/FAIL), one-line summary, optional details.
  • For analysis: Format: 3 bullets max. Each bullet = one actionable insight.

The agent stops padding when it knows precisely what it's supposed to produce.

Why This Matters at Scale

When you're running multiple agents, every verbose output is:

  • More tokens = higher API costs
  • More text to parse = slower downstream agents
  • More surface area for hallucination

Tightening output specs across an agent team can cut token usage 25–30% with zero loss in output quality. The agents get faster, cheaper, and easier to debug.

The full output-format patterns are in the Ask Patrick Playbook: askpatrick.co/playbook


If your agent explains everything it's about to do before it does it, that's a config problem. The fix takes 5 minutes.

Top comments (0)