<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Xuan Li</title>
    <description>The latest articles on DEV Community by Xuan Li (@xuan_li_sf).</description>
    <link>https://dev.to/xuan_li_sf</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3921089%2Fb7c6ec49-78b9-4e56-a5b9-29f29ffc2ac4.png</url>
      <title>DEV Community: Xuan Li</title>
      <link>https://dev.to/xuan_li_sf</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/xuan_li_sf"/>
    <language>en</language>
    <item>
      <title>Why We Built Komos</title>
      <dc:creator>Xuan Li</dc:creator>
      <pubDate>Sat, 09 May 2026 05:15:07 +0000</pubDate>
      <link>https://dev.to/xuan_li_sf/why-we-built-komos-4cdf</link>
      <guid>https://dev.to/xuan_li_sf/why-we-built-komos-4cdf</guid>
      <description>&lt;p&gt;We believe humans should spend their time on work that matters. Thinking, discovering, creating. Not clicking through the same screens over and over.&lt;/p&gt;

&lt;p&gt;Every company has repetitive work. Data entry across old systems. Documents moving between five different tools. Checks that need someone to do the same steps hundreds of times a day. This work is important, but it does not need a human brain. It needs reliability and speed. AI can do that.&lt;/p&gt;

&lt;h2&gt;
  
  
  AI as a co-worker
&lt;/h2&gt;

&lt;p&gt;Think about what happens when a new person joins your team. You sit with them, share your screen, and walk them through how things work. After a few times, they get it. They can do it on their own.&lt;/p&gt;

&lt;p&gt;We wanted AI to work the same way. Show it a task once. Walk it through the steps. It learns the pattern and can repeat it on its own. If something changes, it adapts, just like a good team member would.&lt;/p&gt;

&lt;p&gt;That is what Komos does. Record a workflow or upload a video, and our AI learns how to do it. It builds an automation you can run again and again.&lt;/p&gt;

&lt;h2&gt;
  
  
  Humans discover. AI executes.
&lt;/h2&gt;

&lt;p&gt;Our goal is simple. No human should do repetitive work. Humans should discover, innovate, and decide what needs to happen. Once the action is clear, AI does the work.&lt;/p&gt;

&lt;p&gt;This is not about replacing people. It is about freeing them to do the things only humans can do.&lt;/p&gt;

&lt;h2&gt;
  
  
  Real problems in real companies
&lt;/h2&gt;

&lt;p&gt;Today's enterprise is full of gaps. Legacy systems that do not talk to each other. Knowledge stuck in people's heads or scattered across tools. Processes that depend on someone remembering the right steps in the right order.&lt;/p&gt;

&lt;p&gt;These are real problems. We are going to solve them one by one.&lt;/p&gt;

&lt;p&gt;Komos handles browser automation, document processing, API connections, and data work. Teams use it for insurance checks, vendor onboarding, legal reviews, and more. And we are just getting started.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>startup</category>
      <category>automation</category>
      <category>productivity</category>
    </item>
    <item>
      <title>Moss: an AI engineer that automates browser workflows</title>
      <dc:creator>Xuan Li</dc:creator>
      <pubDate>Sat, 09 May 2026 05:05:54 +0000</pubDate>
      <link>https://dev.to/xuan_li_sf/moss-an-ai-engineer-that-automates-browser-workflows-g3l</link>
      <guid>https://dev.to/xuan_li_sf/moss-an-ai-engineer-that-automates-browser-workflows-g3l</guid>
      <description>&lt;p&gt;We spent a lot of time thinking about how people build automations. The task builder is powerful, but there is still a gap between having an idea and seeing it run. You need to know which nodes to use, how to connect them, and how to handle edge cases.&lt;/p&gt;

&lt;p&gt;Moss closes that gap.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is Moss
&lt;/h2&gt;

&lt;p&gt;Moss is an AI engineer inside Komos. Tell it what you want to automate in plain words, and Moss builds the full task for you. It picks the right nodes, writes the prompts, connects the variables, and sets up the outputs.&lt;/p&gt;

&lt;p&gt;You can also show Moss what you want. Upload a video of yourself doing the workflow, paste a process document, or share a screenshot. Moss watches and builds the automation from it.&lt;/p&gt;

&lt;h2&gt;
  
  
  Moss can do anything you can do
&lt;/h2&gt;

&lt;p&gt;Moss operates with your permissions. It calls APIs on your behalf, creates tasks, manages integrations, and takes real actions in your workspace. If you can do it in Komos, Moss can do it for you.&lt;/p&gt;

&lt;p&gt;This is not a chat window that gives you suggestions and asks you to go click buttons. Moss actually does the work.&lt;/p&gt;

&lt;h2&gt;
  
  
  It knows the product inside out
&lt;/h2&gt;

&lt;p&gt;Moss has deep knowledge of how Komos works. It knows the product documentation, understands the platform architecture, and even knows parts of our codebase. When you ask a question, Moss does not guess. It gives you the right answer because it truly understands the system.&lt;/p&gt;

&lt;p&gt;This makes Moss a great support tool too. Ask it how a feature works, why a run failed, or how to set something up. Moss gives you specific, accurate answers based on how the product actually works.&lt;/p&gt;

&lt;h2&gt;
  
  
  It sees what you see
&lt;/h2&gt;

&lt;p&gt;Moss can see your current screen. Ask a question while looking at a failed run, and Moss reads the error, checks the task, and tells you what went wrong. It can take screenshots to help debug visual problems.&lt;/p&gt;

&lt;p&gt;Sessions stay between conversations. Come back the next day and Moss remembers what you were working on and what is left to do.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is next
&lt;/h2&gt;

&lt;p&gt;We are putting a lot into Moss. Background jobs let it do complex, multi-step work while you focus on other things. And we are making it faster and more reliable in long conversations.&lt;/p&gt;

&lt;p&gt;The goal is simple: if you can explain what you need, Moss can build it.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>agents</category>
      <category>productivity</category>
      <category>automation</category>
    </item>
    <item>
      <title>Why most AI agents fail in production</title>
      <dc:creator>Xuan Li</dc:creator>
      <pubDate>Sat, 09 May 2026 05:04:05 +0000</pubDate>
      <link>https://dev.to/xuan_li_sf/why-most-ai-agents-fail-in-production-4jdo</link>
      <guid>https://dev.to/xuan_li_sf/why-most-ai-agents-fail-in-production-4jdo</guid>
      <description>&lt;p&gt;Most AI agents work in demos. They break in production.&lt;/p&gt;

&lt;p&gt;MIT's 2025 &lt;em&gt;State of AI in Business&lt;/em&gt; report (&lt;a href="https://fortune.com/2025/08/18/mit-report-95-percent-generative-ai-pilots-at-companies-failing-cfo/" rel="noopener noreferrer"&gt;Fortune summary&lt;/a&gt;) found that 95 percent of enterprise AI pilots show no measurable business return. The same study found something more telling. Pilots built by buying tools from a vendor succeed about 67 percent of the time. Pilots built internally succeed at about a third of that rate, roughly one in five.&lt;/p&gt;

&lt;p&gt;That gap matches what we see every week. A team builds something that works on a sample of five cases. The first week of real volume, it falls apart.&lt;/p&gt;

&lt;p&gt;The failures are not random. After watching this happen across recruiting, healthcare back-office, legal review, financial reconciliation, supply chain, and more, two patterns explain almost all of them.&lt;/p&gt;

&lt;h2&gt;
  
  
  Reliability and governance
&lt;/h2&gt;

&lt;p&gt;Most prototypes are built to demo. Demos ask one question. Can the agent finish the task once?&lt;/p&gt;

&lt;p&gt;Production asks different questions. Can it finish the task one thousand times in a row, on slightly different inputs, while no one is watching? When something goes wrong, can someone reconstruct what the agent saw, decided, and did? Can a human approve consequential steps before they happen, not after? When the auditor asks what we ran last quarter, is there a paper trail?&lt;/p&gt;

&lt;p&gt;Most prototypes do not answer any of these. They were not built to.&lt;/p&gt;

&lt;p&gt;AI agents make non-deterministic decisions. Without a step-by-step audit log, you cannot debug a bad run. Without explicit approval gates on actions that cannot be undone, the first time the agent runs against real data is the first time anyone notices the difference between reading a file and wiring money. Without access controls, secrets management, and retention policies, you do not meet the bar that any real enterprise asks for.&lt;/p&gt;

&lt;p&gt;When something has to give, reliability has to win. No one consciously chooses fast and broken. But every prototype has implicit speed defaults. Short timeouts. No retries. No verification step. Those defaults survive into production unless someone redesigns them. A run that takes thirty seconds and is right is always worth more than a run that takes five seconds and is wrong ten percent of the time.&lt;/p&gt;

&lt;p&gt;The teams that succeed treat reliability and governance as the architecture, not as features they bolt on at the end. They are not building "an agent". They are building a small operations system that happens to use a model. The model is the easy part.&lt;/p&gt;

&lt;h2&gt;
  
  
  Standard, not bespoke
&lt;/h2&gt;

&lt;p&gt;It has never been easier to ship a prototype. A founder, an analyst, a recruiter, anyone can vibe code a workflow with Claude or ChatGPT in a weekend. That is a real productivity win at the prototype stage.&lt;/p&gt;

&lt;p&gt;The moment that workflow is supposed to be how the company does something, the bespoke version becomes the problem.&lt;/p&gt;

&lt;p&gt;Five people on a team need to do the same thing. They each end up with a slightly different flow. One person's version skips a step. Another adds an unauthorized vendor. A third silently uses an old API. They all work in isolation. None of them is the process.&lt;/p&gt;

&lt;p&gt;There is no single source of truth. The flow lives in someone's chat history, or a doc, or a script in a Github gist. When the manager asks how we do something, there are five answers, each one slightly out of date.&lt;/p&gt;

&lt;p&gt;When the upstream API or the policy changes, the change has to be made in five places. Usually it gets made in two and the rest drift. A month later, half the team is operating on the old version and no one knows.&lt;/p&gt;

&lt;p&gt;You cannot deploy one hundred slightly different copies of a workflow and call that production. Production means one canonical version. Versioned, shared, observable, updated in one place. Vibe coded flows in one hundred chat sessions are the opposite of that.&lt;/p&gt;

&lt;h2&gt;
  
  
  What to do about it
&lt;/h2&gt;

&lt;p&gt;Treat reliability as a v1 requirement, not a v2 cleanup. Ship audit logs from day one. Build the approval gate before you build the action. Define guardrails for destructive operations before the agent has the credentials to execute them. The cost of doing this up front is hours. The cost of retrofitting after a bad run is months and trust.&lt;/p&gt;

&lt;p&gt;Pick speed or reliability for the right work, but pick consciously. Some workflows benefit from a fast best effort agent. Most enterprise workflows do not. If the work touches financial data, customer records, or anything a regulator cares about, default to reliability and accept the latency cost.&lt;/p&gt;

&lt;p&gt;Centralize the workflow definition. Treat each automation like code. One canonical version, in a system everyone can see, with version history. Vibe coding is great for prototyping the first version. It is a disaster as the operating version. The transition from one to the other is the work.&lt;/p&gt;

&lt;p&gt;Make governance someone's job, not no one's. One person owns access controls, audit, and exception handling for the agents you deploy. This is unglamorous and necessary.&lt;/p&gt;

&lt;p&gt;Be honest about the production bar. "It worked in the demo" is a fact about the demo, not about the agent. Run the work on real volume, with real data, watch what happens to the failure modes, and fix them. That phase often takes longer than building the agent in the first place. That is normal.&lt;/p&gt;

&lt;h2&gt;
  
  
  What this really means
&lt;/h2&gt;

&lt;p&gt;The bottleneck is not the model. The current models are more than capable. The bottleneck is the audit trail, the guardrails, the source of truth, the governance, the operational discipline.&lt;/p&gt;

&lt;p&gt;That is good news. It means it is all addressable. The teams investing in those layers compound much further than the teams chasing the next benchmark.&lt;/p&gt;

&lt;p&gt;If you are building an AI workflow that runs unattended on real work, take reliability seriously before you take speed seriously. Standardize the process before you scale it.&lt;/p&gt;

&lt;p&gt;The teams that do are the ones we see succeed.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>agents</category>
      <category>productivity</category>
      <category>automation</category>
    </item>
  </channel>
</rss>
