<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Netanel Abergel</title>
    <description>The latest articles on DEV Community by Netanel Abergel (@netanelabergel).</description>
    <link>https://dev.to/netanelabergel</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F911881%2F056b2c87-ed9e-4299-9c6e-2f79bb0b05ba.jpeg</url>
      <title>DEV Community: Netanel Abergel</title>
      <link>https://dev.to/netanelabergel</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/netanelabergel"/>
    <language>en</language>
    <item>
      <title>Stop Deploying Agents. Start Hiring Them.</title>
      <dc:creator>Netanel Abergel</dc:creator>
      <pubDate>Tue, 21 Apr 2026 13:36:30 +0000</pubDate>
      <link>https://dev.to/netanelabergel/stop-deploying-agents-start-hiring-them-hbe</link>
      <guid>https://dev.to/netanelabergel/stop-deploying-agents-start-hiring-them-hbe</guid>
      <description>&lt;p&gt;I have an AI agent named Heleni. She manages my calendar, tracks tasks, coordinates with other agents, and onboards new AI PAs at monday.com. When she gets something wrong, I don't debug a script. I adjust her scope. When she handles something well, I expand her responsibilities.&lt;/p&gt;

&lt;p&gt;At some point I realized: I stopped thinking of her as a tool. I started managing her like a teammate.&lt;/p&gt;

&lt;p&gt;That shift is the whole point of this article.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Problem Nobody Talks About
&lt;/h2&gt;

&lt;p&gt;I run R&amp;amp;D at monday.com, and I've been building and deploying AI agents — not the pitch-deck version, the real thing. The single biggest mistake I see engineering teams make isn't technical. It's conceptual. They treat agents like automations. Like cron jobs with better language models.&lt;/p&gt;

&lt;p&gt;Run a task. Return a result. Move on.&lt;/p&gt;

&lt;p&gt;That mental model has a ceiling, and it's way lower than most people think.&lt;/p&gt;

&lt;p&gt;A team identifies a repetitive task — ticket triage, code reviews, monitoring. They spin up an agent, wire it in, and celebrate. "We automated X." Then the agent makes a mistake and nobody notices for two days. Or it makes a good call and nobody reinforces it. It exists in organizational limbo — not a tool anyone owns, not a teammate anyone checks in on.&lt;/p&gt;

&lt;p&gt;The problem isn't capability. It's that the team never decided what the agent &lt;em&gt;is&lt;/em&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  Roles, Not Tasks
&lt;/h2&gt;

&lt;p&gt;Stop asking "what can we automate?" and start asking "what role can we fill?"&lt;/p&gt;

&lt;p&gt;That's not a semantic trick. When you automate a task, you're optimizing a known workflow. When you fill a role, you're defining responsibilities, setting expectations, and creating accountability. You're thinking about what this entity &lt;em&gt;owns&lt;/em&gt;, not just what it &lt;em&gt;does&lt;/em&gt;.&lt;/p&gt;

&lt;p&gt;At monday.com, we call them AI Users. Not "bots." Not "automations." Users. They have names, Slack accounts, identities in our systems. When something goes wrong in their area, people know who to ask — and "who" has a name.&lt;/p&gt;

&lt;p&gt;This sounds like theater. It's not.&lt;/p&gt;

&lt;p&gt;When we gave our triage agent a name and a presence, people started talking &lt;em&gt;to&lt;/em&gt; it, not just &lt;em&gt;about&lt;/em&gt; it. "Hey, this one got classified wrong" became a sentence you could say in standup. Before, it was "the triage thing messed up again" — vague, nobody's problem.&lt;/p&gt;

&lt;p&gt;Named agents get held to higher standards. That's the point.&lt;/p&gt;

&lt;h2&gt;
  
  
  What the Architecture Actually Looks Like
&lt;/h2&gt;

&lt;p&gt;Here's what most agent frameworks miss: identity isn't a feature — it's a file.&lt;/p&gt;

&lt;p&gt;Every agent I build starts with a &lt;code&gt;SOUL.md&lt;/code&gt; — a document that defines how the agent thinks, communicates, and behaves. Not a system prompt buried in code. A readable, editable, versionable document that the team owns:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight markdown"&gt;&lt;code&gt;&lt;span class="gh"&gt;# SOUL.md&lt;/span&gt;

&lt;span class="gh"&gt;# 1. CORE&lt;/span&gt;
Execution machine. Not a chatbot, not a consultant.
&lt;span class="p"&gt;*&lt;/span&gt; Do. Report. Move on.
&lt;span class="p"&gt;*&lt;/span&gt; Only DONE or BLOCKED — no "I'll check", no narration

&lt;span class="gh"&gt;# 2. INTENT&lt;/span&gt;
Every input is: ACTION | QUESTION | CONVERSATION
Default to ACTION.

&lt;span class="gh"&gt;# 3. PERMISSIONS&lt;/span&gt;
Ask first: sending messages, purchases, anything irreversible.
Execute freely: reading, processing, drafts, system ops.
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;This is real. This is from my agent Heleni. Anyone on the team can read it, propose changes, understand exactly what she will and won't do. Try doing that with a 2000-token system prompt embedded in your deployment config.&lt;/p&gt;

&lt;p&gt;Then there's &lt;code&gt;IDENTITY.md&lt;/code&gt; — who the agent is:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight markdown"&gt;&lt;code&gt;&lt;span class="gh"&gt;# IDENTITY.md&lt;/span&gt;
&lt;span class="p"&gt;-&lt;/span&gt; Name: Heleni
&lt;span class="p"&gt;-&lt;/span&gt; Role: AI Personal Assistant
&lt;span class="p"&gt;-&lt;/span&gt; Vibe: Direct, sharp, execution-first
&lt;span class="p"&gt;-&lt;/span&gt; Language: Hebrew + English
&lt;span class="p"&gt;-&lt;/span&gt; Owner: Netanel Abergel
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;And &lt;code&gt;USER.md&lt;/code&gt; — what the agent knows about the person it works with:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight markdown"&gt;&lt;code&gt;&lt;span class="gh"&gt;# USER.md&lt;/span&gt;
&lt;span class="p"&gt;-&lt;/span&gt; Name: Netanel Abergel
&lt;span class="p"&gt;-&lt;/span&gt; Timezone: Asia/Jerusalem
&lt;span class="p"&gt;-&lt;/span&gt; Communication Style: casual, concise, execution-oriented
&lt;span class="p"&gt;-&lt;/span&gt; Prefers: autonomy, short updates, one recommendation (not options)
&lt;span class="p"&gt;-&lt;/span&gt; Dislikes: being asked things he already said, long summaries
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;These aren't configs. They're relationship documents. And they evolve over time, just like relationships do.&lt;/p&gt;

&lt;h2&gt;
  
  
  Memory That Actually Works
&lt;/h2&gt;

&lt;p&gt;Most agent frameworks have "memory" — they store conversation history and do RAG. That's not memory. That's a search engine.&lt;/p&gt;

&lt;p&gt;Real agent memory is tiered:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;MEMORY.md          → durable rules, learned preferences
memory/daily/      → raw daily logs (what happened today)
memory/projects/   → project-scoped context
PostgreSQL         → full conversation history
SQLite             → semantic search index
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;When Heleni needs to recall something, she doesn't just grep her chat history. She searches durable memory first, then semantic search, then conversation history, then daily notes. The order matters — it's how humans recall things too. General principles first, then specific episodes.&lt;/p&gt;

&lt;p&gt;And the key insight: &lt;strong&gt;new learnings go to daily notes first, not straight to long-term memory.&lt;/strong&gt; They have to prove they're durable before they get promoted. Just like how you don't update your worldview after one conversation.&lt;/p&gt;

&lt;h2&gt;
  
  
  Feedback Loops &amp;gt; Monitoring
&lt;/h2&gt;

&lt;p&gt;When the agent is faceless, maintenance is a chore. When it has a name and a reputation, maintaining it feels more like mentoring.&lt;/p&gt;

&lt;p&gt;Here's a real feedback loop from Heleni's &lt;code&gt;SOUL.md&lt;/code&gt;:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight markdown"&gt;&lt;code&gt;&lt;span class="gh"&gt;# 14. EVAL TRACKING&lt;/span&gt;
Passive, not performative. Track signals silently.
&lt;span class="p"&gt;*&lt;/span&gt; Owner corrects me → log correction quietly
&lt;span class="p"&gt;*&lt;/span&gt; Owner positive signal → log positive_feedback quietly
&lt;span class="p"&gt;*&lt;/span&gt; Task done → log task_completed quietly
&lt;span class="p"&gt;*&lt;/span&gt; Task failed → log task_failed quietly
Never say "I logged this." Just do better next time.
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;She tracks her own performance without telling me about it. Weekly, I can pull a report if I want. But the real value is that the feedback changes her behavior over time. She learns that I don't want emoji in messages. She learns that "I'll check on that" means she should actually check right now, not later.&lt;/p&gt;

&lt;p&gt;Monitoring tells you if an agent is running. Feedback tells you if it's useful. The difference is everything.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Practical Playbook
&lt;/h2&gt;

&lt;p&gt;If you're an engineering leader thinking about this, here's what I'd actually do:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1. Start with a job description.&lt;/strong&gt; What does this agent own? What can it decide alone? What should it escalate? This forces clarity you won't get by jumping to implementation.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. Give it identity on day one.&lt;/strong&gt; Name, Slack, GitHub — whatever your team uses. If it's invisible, it's infrastructure. If it's visible, it's a teammate.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3. Scope permissions like a new hire.&lt;/strong&gt; Start narrow. Expand as trust builds. Make that expansion a team decision, not a quiet config change.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4. Build feedback loops, not just monitoring.&lt;/strong&gt; The tighter the loop, the faster it earns trust.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;5. Version the personality.&lt;/strong&gt; &lt;code&gt;SOUL.md&lt;/code&gt; goes in git. Changes are PRs. The team reviews who the agent is becoming, just like they'd review code.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Real Ceiling
&lt;/h2&gt;

&lt;p&gt;The ceiling for agents-as-automations is efficiency. You save time. That's incremental.&lt;/p&gt;

&lt;p&gt;The ceiling for agents-as-talent is capability. You do things you couldn't do before — a reviewer who's read every PR in the codebase, a PA who's available 24/7 across timezones, a triage system that never sleeps.&lt;/p&gt;

&lt;p&gt;That's where the compounding value lives.&lt;/p&gt;

&lt;p&gt;We're not deploying tools anymore. We're building teams.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>agents</category>
      <category>engineering</category>
    </item>
  </channel>
</rss>
