<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Seth Rose</title>
    <description>The latest articles on DEV Community by Seth Rose (@thesethrose).</description>
    <link>https://dev.to/thesethrose</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3432607%2F8fcc650b-c00f-49f0-ae9b-01ed3c52bacb.jpeg</url>
      <title>DEV Community: Seth Rose</title>
      <link>https://dev.to/thesethrose</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/thesethrose"/>
    <language>en</language>
    <item>
      <title>The Jobs AI Can’t Touch (Yet): Why Some Roles Are Safe from Automation</title>
      <dc:creator>Seth Rose</dc:creator>
      <pubDate>Thu, 14 Aug 2025 20:28:32 +0000</pubDate>
      <link>https://dev.to/thesethrose/the-jobs-ai-cant-touch-yet-why-some-roles-are-safe-from-automation-192f</link>
      <guid>https://dev.to/thesethrose/the-jobs-ai-cant-touch-yet-why-some-roles-are-safe-from-automation-192f</guid>
      <description>&lt;p&gt;AI has been chewing through industries like Pac-Man lately, but not every job is on the menu.&lt;/p&gt;

&lt;p&gt;While software engineers and call center reps feel the heat, some roles like dredge operators, bridge tenders, and water treatment plant operators are surprisingly safe.&lt;/p&gt;

&lt;p&gt;This article breaks down why certain jobs are &lt;em&gt;automation-resistant&lt;/em&gt; and what it teaches us about the limits of AI. By the end, you’ll know how to spot safe zones in the workforce and why your industry might not be next in line for the robot takeover.&lt;/p&gt;

&lt;p&gt;After reading this article, you’ll:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Understand why physical-world complexity slows automation.&lt;/li&gt;
&lt;li&gt;See real-world examples of jobs AI struggles to replace.&lt;/li&gt;
&lt;li&gt;Learn how developers can spot opportunities &lt;em&gt;and&lt;/em&gt; limits in automation.&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  The Jobs AI Can’t Replace Easily
&lt;/h2&gt;

&lt;p&gt;When Microsoft researchers analyzed 20,000 Copilot chats to predict automation impact, some roles were barely touched by AI.&lt;/p&gt;

&lt;p&gt;Think dredge operators, pile driver operators, or bridge and lock tenders. Why? &lt;/p&gt;

&lt;p&gt;Because these jobs:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Require physical interaction with the environment.&lt;/li&gt;
&lt;li&gt;Operate in unpredictable or dangerous conditions.&lt;/li&gt;
&lt;li&gt;Have high safety and legal oversight.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;💡 &lt;strong&gt;Example:&lt;/strong&gt; A dredge operator navigates changing water currents, unexpected debris, and machinery maintenance in real-time. The variables are too many and too costly to automate without a massive R&amp;amp;D investment.&lt;/p&gt;




&lt;h2&gt;
  
  
  Why AI Struggles Here
&lt;/h2&gt;

&lt;p&gt;AI thrives in predictable, digital environments. In contrast, these jobs involve:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;High variability:&lt;/strong&gt; Nature and physical systems rarely behave in a neat, repeatable way.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Low automation ROI:&lt;/strong&gt; The cost of developing, deploying, and maintaining robots for these jobs often outweighs the savings.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Regulation &amp;amp; liability:&lt;/strong&gt; If an autonomous bridge tender fails, lives are at risk. That’s not a risk companies or regulators rush to take.&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  What This Means for Developers
&lt;/h2&gt;

&lt;p&gt;If you’re building AI tools, here’s the takeaway: not all problems are worth automating right now. Instead of chasing every possible job replacement, look for areas where:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;The environment is controlled (like warehouses or server rooms).&lt;/li&gt;
&lt;li&gt;Tasks are repetitive and rule-based.&lt;/li&gt;
&lt;li&gt;Stakeholders are open to rapid experimentation.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;💡 &lt;strong&gt;Example for Devs:&lt;/strong&gt; Instead of trying to build a fully autonomous bridge control system, focus on decision-support tools like predictive maintenance dashboards that assist human operators without replacing them.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Hidden Adoption Gap
&lt;/h2&gt;

&lt;p&gt;Even when automation is technically possible, adoption lags behind the hype. Factors like union negotiations, insurance costs, and slow procurement cycles can stretch timelines by years.&lt;/p&gt;

&lt;p&gt;This gap is your opportunity. If you understand the &lt;em&gt;real&lt;/em&gt; pace of adoption, you can build products that solve immediate, human-in-the-loop problems while the market catches up.&lt;/p&gt;




&lt;h2&gt;
  
  
  My Take
&lt;/h2&gt;

&lt;p&gt;We tend to think of AI as an unstoppable force mowing down every job in its path. The reality is messier and more interesting. Some work remains human not because AI isn’t smart enough, but because the real world is a chaotic, costly, and highly regulated place.&lt;/p&gt;

&lt;p&gt;That’s good news for developers. It means there’s still massive room for human-centered tools that enhance, not replace, the people in these roles. In the short term, hybrid AI+human workflows are often the smartest bet.&lt;/p&gt;




&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;Not every job is at risk of immediate automation. Understanding why can make you a better developer, investor, or career planner. Focus on problems where AI fits the environment, economics, and risk profile, and don’t underestimate the staying power of human skill.&lt;/p&gt;

&lt;p&gt;If you’re building AI products, aim for tools that &lt;em&gt;assist&lt;/em&gt; in complex environments instead of replacing them. The AI gold rush isn’t just about who can automate first; it’s about who can build what people will actually trust and adopt.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Based on &lt;a href="https://www.businessinsider.com/ai-jobs-overlap-affect-research-2025-7" rel="noopener noreferrer"&gt;Microsoft’s analysis of 200,000 Copilot chatbot conversations&lt;/a&gt;, which identified jobs most and least affected by generative AI, and broader discussions about the real-world pace of AI adoption.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>automation</category>
      <category>futureofwork</category>
    </item>
    <item>
      <title>Why Your AI Agent Is Failing (and How to Fix It)</title>
      <dc:creator>Seth Rose</dc:creator>
      <pubDate>Wed, 13 Aug 2025 21:10:25 +0000</pubDate>
      <link>https://dev.to/thesethrose/why-your-ai-agent-is-failing-and-how-to-fix-it-4ne2</link>
      <guid>https://dev.to/thesethrose/why-your-ai-agent-is-failing-and-how-to-fix-it-4ne2</guid>
      <description>&lt;p&gt;Most AI agent failures don’t happen because the model isn’t “smart enough.” They happen because the system around them wasn’t built to succeed.&lt;/p&gt;

&lt;p&gt;In my work building LLM agents, here’s what I encounter most often:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Unreliable prompt architecture&lt;/strong&gt;&lt;br&gt;
Agents usually depend on multiple prompts for planning, memory and tool use. Even minor formatting shifts can break robustness. You may need to test prompts systematically, tune them, or generate variants using the model itself.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Weak or missing evaluation strategy&lt;/strong&gt;&lt;br&gt;
Having no way to measure progress is like flying blind. Effective evals should go beyond end-to-end success. They must test components like tool calls, reasoning chains, and completion accuracy. Component level tracing helps isolate where things break.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Lack of safety and adversarial defenses&lt;/strong&gt;&lt;br&gt;
Prompt injection and memory poisoning are real threats: attacks often succeed in benchmarks. Basic defense prompts alone aren’t enough. Build robust safety audits and testing frameworks.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Poor system design or ambiguous spec&lt;/strong&gt;&lt;br&gt;
In multi-agent setups, many failures come from vague task specs or unclear role handoffs. Misaligned workflows and weak termination logic often sabotage execution.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;No human-in-the-loop or judge feedback loop&lt;/strong&gt;&lt;br&gt;
Automated “judge” LLM evals can drift, if you train one black box to evaluate another, you’re stacking assumptions. The same works for security. The best pipelines mix automated scoring with occasional human reviews to catch what machines miss.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Tool invocation confusion&lt;/strong&gt;&lt;br&gt;
If your agent can call APIs or plugins, the interface needs to be rock solid. JSON schemas or function signatures work much better than loose natural-language descriptions, you want predictable invocation formats.&lt;/p&gt;

&lt;h2&gt;
  
  
  How to Fix It
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Start by drawing a clear roadmap: what's the end goal, and which component is responsible for which part?
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Build robustness one prompt at a time&lt;/strong&gt;&lt;br&gt;
Treat prompt design like a polished API. Use clear instructions, delimiters, role definitions, and log everything. Track variants over time to see what actually works.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Run component-level evals&lt;/strong&gt;&lt;br&gt;
Create traces for tool usage, step-by-step reasoning, and task completion. Test both happy paths and edge cases so you can pinpoint failures.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Add real safety checks&lt;/strong&gt;&lt;br&gt;
Simulate prompt injections and memory corruption. Run safety benchmarks and audit your agent’s attack surface regularly.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Clarify spec and roles&lt;/strong&gt;&lt;br&gt;
Write crisp specs for agents. Define “this agent does X, then passes to that agent,” and make termination conditions explicit. Use a failure taxonomy to guide your audit.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Hybrid evaluation loops&lt;/strong&gt;&lt;br&gt;
Use “LLM-as-a-judge” to scale reviews but spot-check with humans. Regularly calibrate the judge to stay accurate.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Standardize tool calls&lt;/strong&gt;&lt;br&gt;
Define tool schemas or function signatures, and embed that in context so agents know exactly how to call them. This avoids parsing confusion and silent failures.&lt;/p&gt;

&lt;h3&gt;
  
  
  Quick checklist
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Define end-to-end success criteria and component metrics.
&lt;/li&gt;
&lt;li&gt;Freeze a prompt baseline and iterate with controlled experiments.
&lt;/li&gt;
&lt;li&gt;Add staged safety tests before deployment.
&lt;/li&gt;
&lt;li&gt;Require a human sign-off for high-risk decisions.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Have you run into any of these pain points while building agents?&lt;/p&gt;

&lt;p&gt;Curious to hear if you’ve tried component-level evals or a prompt robustness test and whether it revealed something wild.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>agentaichallenge</category>
      <category>promptengineering</category>
      <category>evals</category>
    </item>
  </channel>
</rss>
