<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: yan yan</title>
    <description>The latest articles on DEV Community by yan yan (@yan_yan_096c6a71b7657ec65).</description>
    <link>https://dev.to/yan_yan_096c6a71b7657ec65</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3929761%2Fd8f44b99-c38e-4c70-8c1d-de8c33b1065a.png</url>
      <title>DEV Community: yan yan</title>
      <link>https://dev.to/yan_yan_096c6a71b7657ec65</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/yan_yan_096c6a71b7657ec65"/>
    <language>en</language>
    <item>
      <title>Build Your Own AI Writing Pipeline: From Idea to Published in 30 Minutes</title>
      <dc:creator>yan yan</dc:creator>
      <pubDate>Wed, 13 May 2026 16:50:28 +0000</pubDate>
      <link>https://dev.to/yan_yan_096c6a71b7657ec65/build-your-own-ai-writing-pipeline-from-idea-to-published-in-30-minutes-cgl</link>
      <guid>https://dev.to/yan_yan_096c6a71b7657ec65/build-your-own-ai-writing-pipeline-from-idea-to-published-in-30-minutes-cgl</guid>
      <description>&lt;p&gt;Liquid syntax error: Unknown tag 'endraw'&lt;/p&gt;
</description>
      <category>ai</category>
      <category>python</category>
      <category>tutorial</category>
      <category>productivity</category>
    </item>
    <item>
      <title>How I Built an AI Content Machine That Publishes 40 Articles a Month</title>
      <dc:creator>yan yan</dc:creator>
      <pubDate>Wed, 13 May 2026 16:48:35 +0000</pubDate>
      <link>https://dev.to/yan_yan_096c6a71b7657ec65/how-i-built-an-ai-content-machine-that-publishes-40-articles-a-month-1ek3</link>
      <guid>https://dev.to/yan_yan_096c6a71b7657ec65/how-i-built-an-ai-content-machine-that-publishes-40-articles-a-month-1ek3</guid>
      <description>&lt;h1&gt;
  
  
  I Built an AI Content Machine That Runs 24/7. Here Is the Blueprint.
&lt;/h1&gt;

&lt;h2&gt;
  
  
  No team. No budget. Just one laptop, three tools, and a system that never sleeps.
&lt;/h2&gt;




&lt;p&gt;Three months ago, I was spending 12 hours a week writing content. Blog posts. Newsletter. Social media. The grind was real, and the results were mediocre.&lt;/p&gt;

&lt;p&gt;Today? I spend 30 minutes a week on content. The machine does the rest.&lt;/p&gt;

&lt;p&gt;Here is exactly how I built it — no fluff, no newsletter pitch, just the tools, the workflow, and the mistakes I made along the way.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Problem Nobody Talks About
&lt;/h2&gt;

&lt;p&gt;Most people think the hard part of content creation is writing. It is not.&lt;/p&gt;

&lt;p&gt;The hard part is everything else:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Coming up with ideas worth writing about&lt;/li&gt;
&lt;li&gt;Researching and organizing information&lt;/li&gt;
&lt;li&gt;Editing for clarity and tone&lt;/li&gt;
&lt;li&gt;Formatting for different platforms&lt;/li&gt;
&lt;li&gt;Publishing consistently even when you do not feel like it&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Writing is maybe 20% of the work. The other 80% is the machine around it — the system that turns raw thoughts into published content without burning you out.&lt;/p&gt;

&lt;p&gt;Most "AI writing" advice stops at "use ChatGPT to generate outlines." That is like stopping at step one of building a car and saying "I have wheels now."&lt;/p&gt;

&lt;p&gt;Here is the full blueprint.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Architecture
&lt;/h2&gt;

&lt;p&gt;My content machine has four layers:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Ideas --&amp;gt; Research --&amp;gt; Draft --&amp;gt; Polish --&amp;gt; Publish
  ^                                      |
  |                                      v
  +------- Feedback Loop -------+
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Each layer is automated differently. The key insight: &lt;strong&gt;AI is not one tool. It is a pipeline of specialized tools, each doing one thing well.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Here is the stack:&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Layer&lt;/th&gt;
&lt;th&gt;Tool&lt;/th&gt;
&lt;th&gt;Why&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Ideas&lt;/td&gt;
&lt;td&gt;Notion + Custom AI trigger&lt;/td&gt;
&lt;td&gt;Generates 50 ideas from 5 seeds&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Research&lt;/td&gt;
&lt;td&gt;Perplexity API&lt;/td&gt;
&lt;td&gt;Gathers sources, data, quotes&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Draft&lt;/td&gt;
&lt;td&gt;Claude API (long-form)&lt;/td&gt;
&lt;td&gt;Writes the first draft&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Polish&lt;/td&gt;
&lt;td&gt;Grammarly + Manual review&lt;/td&gt;
&lt;td&gt;Fixes tone, flow, facts&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Publish&lt;/td&gt;
&lt;td&gt;Custom script + APIs&lt;/td&gt;
&lt;td&gt;Formats and posts to all platforms&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;




&lt;h2&gt;
  
  
  Layer 1: The Idea Engine
&lt;/h2&gt;

&lt;p&gt;This was my first breakthrough. I stopped trying to "think of ideas" and built a machine to do it.&lt;/p&gt;

&lt;p&gt;The system is dead simple:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Every Monday, I write down 5 &lt;strong&gt;seed topics&lt;/strong&gt; — broad areas I want to cover this week. For example: "remote work productivity," "AI ethics," "freelancing tools."&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;I feed those seeds into an AI prompt that generates 10 article angles per seed. The prompt forces specificity:&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;blockquote&gt;
&lt;p&gt;"For each topic, generate 10 article ideas. Each must include: (1) a specific reader problem, (2) a counterintuitive claim, (3) a concrete example or data point. No generic listicles. No ultimate guides."&lt;/p&gt;
&lt;/blockquote&gt;

&lt;ol&gt;
&lt;li&gt;Out of 50 generated ideas, I pick the top 5 based on a scoring system:

&lt;ul&gt;
&lt;li&gt;Is this actually useful? (0-3 pts)&lt;/li&gt;
&lt;li&gt;Can I write it with authority? (0-3 pts)&lt;/li&gt;
&lt;li&gt;Would I click this headline? (0-3 pts)&lt;/li&gt;
&lt;li&gt;Is anyone else writing this exact article? (-5 pts if yes)&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;I spend exactly 15 minutes on this. The rest of the week is execution.&lt;/p&gt;




&lt;h2&gt;
  
  
  Layer 2: Research Without the Rabbit Hole
&lt;/h2&gt;

&lt;p&gt;Research used to eat half my writing time. I would start Googling one thing, fall into a Wikipedia spiral, and emerge two hours later with 47 open tabs and zero paragraphs written.&lt;/p&gt;

&lt;p&gt;Now I use a structured research prompt fed to Perplexity API:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Find 3-5 credible sources (academic papers, official docs, first-party data — never Medium posts)&lt;/li&gt;
&lt;li&gt;Extract 2-3 key statistics or data points&lt;/li&gt;
&lt;li&gt;Find one counter-argument or opposing view&lt;/li&gt;
&lt;li&gt;Include source URLs&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This gives me a research brief in under 60 seconds. I review it, check the sources, and move on.&lt;/p&gt;

&lt;p&gt;The key rule: &lt;strong&gt;research is for facts, not for inspiration.&lt;/strong&gt; If I do not have enough to start writing after this step, the topic is not ready. Skip it, come back later.&lt;/p&gt;




&lt;h2&gt;
  
  
  Layer 3: The Drafting Pipeline
&lt;/h2&gt;

&lt;p&gt;This is where most people make mistakes. They either:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Use AI to write the whole thing and publish garbage&lt;/li&gt;
&lt;li&gt;Refuse to use AI at all and stay stuck at 2 articles per month&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The right approach is a &lt;strong&gt;drafting pipeline&lt;/strong&gt;: AI writes the skeleton and the muscle. You add the heart.&lt;/p&gt;

&lt;p&gt;My drafting prompt mandates:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Opening&lt;/strong&gt;: Start with a personal anecdote or surprising stat. No "In today is digital age" openings. Ever.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Tone&lt;/strong&gt;: Direct, opinionated, slightly informal. Write like you are explaining something to a smart friend at a bar.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Structure&lt;/strong&gt;: Problem --&amp;gt; Why it matters --&amp;gt; How to fix it --&amp;gt; Concrete steps --&amp;gt; Results/Call to action.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Length&lt;/strong&gt;: 1500-2000 words.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Rules&lt;/strong&gt;: No fluff adjectives. Every paragraph must either teach something or advance the argument. No exceptions.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The output is about 70% there. Good structure, decent flow, but it lacks personality. That is where I come in.&lt;/p&gt;




&lt;h2&gt;
  
  
  Layer 4: The Human Filter
&lt;/h2&gt;

&lt;p&gt;I spend exactly 20 minutes per article on the human edit. Here is my checklist:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Voice injection&lt;/strong&gt;: Rewrite the first paragraph entirely in your voice. This sets the tone for the whole piece.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Example swap&lt;/strong&gt;: Replace generic AI examples with specific personal stories. "A study found that..." becomes "Last month, a client asked me..."&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Contrarian check&lt;/strong&gt;: Does this article say anything surprising? If not, add one paragraph that challenges the reader is assumptions.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Fact verification&lt;/strong&gt;: Spot-check every statistic and link. AI hallucinates. Always verify.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;That is it. 20 minutes. Then it goes to Grammarly for the final polish pass, and from there to the publishing pipeline.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Publishing Engine (The Part Nobody Talks About)
&lt;/h2&gt;

&lt;p&gt;This is where 90% of people quit. Writing is done, but publishing across platforms is a nightmare of formatting, image sizing, and copy-pasting.&lt;/p&gt;

&lt;p&gt;I wrote a lightweight Python script that:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Takes the final markdown file&lt;/li&gt;
&lt;li&gt;Formats it for each platform (Medium HTML, WordPress Gutenberg, Substack, LinkedIn)&lt;/li&gt;
&lt;li&gt;Pushes it to the right API&lt;/li&gt;
&lt;li&gt;Logs the post URL&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;The key part — posting to Medium:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;&lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="n"&gt;requests&lt;/span&gt;

&lt;span class="k"&gt;def&lt;/span&gt; &lt;span class="nf"&gt;publish_to_medium&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;article&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;token&lt;/span&gt;&lt;span class="p"&gt;):&lt;/span&gt;
    &lt;span class="n"&gt;headers&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="p"&gt;{&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;Authorization&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;Bearer &lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt; &lt;span class="o"&gt;+&lt;/span&gt; &lt;span class="n"&gt;token&lt;/span&gt;&lt;span class="p"&gt;}&lt;/span&gt;
    &lt;span class="n"&gt;me&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;requests&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nf"&gt;get&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;
        &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;https://api.medium.com/v1/me&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
        &lt;span class="n"&gt;headers&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="n"&gt;headers&lt;/span&gt;
    &lt;span class="p"&gt;).&lt;/span&gt;&lt;span class="nf"&gt;json&lt;/span&gt;&lt;span class="p"&gt;()&lt;/span&gt;
    &lt;span class="n"&gt;payload&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="p"&gt;{&lt;/span&gt;
        &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;title&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="n"&gt;article&lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;title&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;],&lt;/span&gt;
        &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;contentFormat&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;markdown&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
        &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;content&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="n"&gt;article&lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;body&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;],&lt;/span&gt;
        &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;tags&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="n"&gt;article&lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;tags&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;],&lt;/span&gt;
        &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;publishStatus&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;draft&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;
    &lt;span class="p"&gt;}&lt;/span&gt;
    &lt;span class="n"&gt;resp&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;requests&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nf"&gt;post&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;
        &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;https://api.medium.com/v1/users/&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;
        &lt;span class="o"&gt;+&lt;/span&gt; &lt;span class="n"&gt;me&lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;data&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;][&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;id&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;]&lt;/span&gt; &lt;span class="o"&gt;+&lt;/span&gt; &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;/posts&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
        &lt;span class="n"&gt;headers&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="n"&gt;headers&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
        &lt;span class="n"&gt;json&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="n"&gt;payload&lt;/span&gt;
    &lt;span class="p"&gt;)&lt;/span&gt;
    &lt;span class="k"&gt;return&lt;/span&gt; &lt;span class="n"&gt;resp&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nf"&gt;json&lt;/span&gt;&lt;span class="p"&gt;()&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;The whole pipeline — from idea to published draft on 3 platforms — takes me about 30 minutes per article.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Results
&lt;/h2&gt;

&lt;p&gt;After three months of running this system:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Output&lt;/strong&gt;: 12 articles/month --&amp;gt; 40 articles/month&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Total read time&lt;/strong&gt;: 1,200 minutes/month --&amp;gt; 8,700 minutes/month&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Subscribers&lt;/strong&gt;: 340 --&amp;gt; 2,100&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Time spent writing&lt;/strong&gt;: 48 hours/month --&amp;gt; 8 hours/month&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;And here is the part that surprised me: &lt;strong&gt;quality went up, not down.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;When I was doing everything manually, I was rushing. Skipping research. Publishing first drafts. The machine handles the grunt work so I can focus on what actually matters: the thinking, the voice, the human insight.&lt;/p&gt;




&lt;h2&gt;
  
  
  What I Would Do Differently
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;1. Start with a smaller scope.&lt;/strong&gt; My first version tried to automate everything at once. It broke. Start with one layer, get it working, then add the next.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. Log everything.&lt;/strong&gt; I did not track which prompts worked best for the first month. That was stupid. Now every article has metadata: prompt version, research sources, edit time, performance. A/B test your prompts like you would ads.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3. The voice is the moat.&lt;/strong&gt; Anyone can generate AI content. Nobody can replicate your specific perspective and experiences. The human edit is not where you save time — it is where you create value.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Blueprint (Steal This)
&lt;/h2&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;STEP 1: Idea generation (Monday, 15 min)
  - Feed 5 seed topics into AI prompt
  - Score and select top 5 article ideas

STEP 2: Research sprint (Monday, 30 min)
  - Run all 5 ideas through research prompt
  - Review and save research briefs

STEP 3: Draft (Tue-Wed, 10 min/article)
  - Feed research brief + style guide to AI
  - Generate first drafts for all 5 articles

STEP 4: Human edit (Wed-Thu, 20 min/article)
  - Voice injection, example swap, fact verify

STEP 5: Polish and publish (Friday, 30 min)
  - Grammarly pass
  - Run publishing script
  - Schedule social posts for the week

TOTAL: ~4 hours/week for 5-7 polished articles
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;






&lt;h2&gt;
  
  
  The One Rule
&lt;/h2&gt;

&lt;p&gt;If you take nothing else from this article, take this:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The goal is not to replace yourself with AI. The goal is to replace the parts of your workflow that do not require you.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Your taste. Your experience. Your weird opinions. Those are irreplaceable — and they are exactly what make your content worth reading. Everything else? Automate it.&lt;/p&gt;




&lt;p&gt;If you found this useful, follow for more on building automated systems that let you do more with less. No hustle culture. Just practical engineering applied to creative work.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>productivity</category>
      <category>programming</category>
      <category>webdev</category>
    </item>
    <item>
      <title>I Tested 5 AI Coding Tools on Real Work. Here Are the Results.</title>
      <dc:creator>yan yan</dc:creator>
      <pubDate>Wed, 13 May 2026 16:47:28 +0000</pubDate>
      <link>https://dev.to/yan_yan_096c6a71b7657ec65/i-tested-5-ai-coding-tools-on-real-work-here-are-the-results-4nik</link>
      <guid>https://dev.to/yan_yan_096c6a71b7657ec65/i-tested-5-ai-coding-tools-on-real-work-here-are-the-results-4nik</guid>
      <description>&lt;h1&gt;
  
  
  I Tested 5 AI Coding Tools on Real Work. Here Are the Results.
&lt;/h1&gt;

&lt;blockquote&gt;
&lt;p&gt;I gave Copilot, Cursor, Claude Code, Windsurf, and Aider the same 3 real tasks. The results were not even close.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;AI coding tools are everywhere. GitHub Copilot. Cursor. Claude Code. Windsurf. Aider. Every week there is a new one, and every review says "this tool changed my life."&lt;/p&gt;

&lt;p&gt;I don't trust those reviews. Most test on toy problems — a todo app, sorting an array, fetching from an API. That is not how real software works.&lt;/p&gt;

&lt;p&gt;So I designed a real-world benchmark. Three tasks pulled from my actual work. Not contrived. Not simplified. The same mess you deal with every day.&lt;/p&gt;

&lt;p&gt;Here are the results.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Test Setup
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;The tasks:&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Legacy refactor&lt;/strong&gt;: A 400-line Python script with no tests, no types, and a known bug. Add type hints, write tests, and fix the bug without breaking anything else.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Greenfield feature&lt;/strong&gt;: Build a real-time data pipeline with WebSocket ingestion, transformation, and PostgreSQL writes. Must handle reconnection, backpressure, and schema evolution.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Debug mystery&lt;/strong&gt;: A Node.js service randomly returns 502 errors under load. No error messages. Been open for 2 weeks. Find it.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;The contestants:&lt;/strong&gt;&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Tool&lt;/th&gt;
&lt;th&gt;Type&lt;/th&gt;
&lt;th&gt;Pricing&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;GitHub Copilot&lt;/td&gt;
&lt;td&gt;VS Code extension&lt;/td&gt;
&lt;td&gt;$10/mo&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Cursor&lt;/td&gt;
&lt;td&gt;AI-native IDE&lt;/td&gt;
&lt;td&gt;$20/mo&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Claude Code&lt;/td&gt;
&lt;td&gt;CLI agent&lt;/td&gt;
&lt;td&gt;API pay-per-use&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Windsurf&lt;/td&gt;
&lt;td&gt;AI IDE&lt;/td&gt;
&lt;td&gt;$15/mo&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Aider&lt;/td&gt;
&lt;td&gt;Open-source CLI&lt;/td&gt;
&lt;td&gt;Free (API cost only)&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Scoring (1-10):&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Accuracy&lt;/strong&gt;: Did it produce correct, working code?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Context awareness&lt;/strong&gt;: Did it understand the existing codebase?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Autonomy&lt;/strong&gt;: How much did I have to hand-hold?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Speed&lt;/strong&gt;: Time from prompt to working solution.&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Task 1: Legacy Refactor
&lt;/h2&gt;

&lt;p&gt;The script processes CSV files from an IoT sensor fleet. 400 lines. Zero comments. Variable names like &lt;code&gt;x&lt;/code&gt;, &lt;code&gt;tmp&lt;/code&gt;, and &lt;code&gt;stuff&lt;/code&gt;. Been in production for 2 years. Nobody wants to touch it.&lt;/p&gt;

&lt;p&gt;The known bug: on files larger than 10MB, the script silently drops the last batch of rows.&lt;/p&gt;

&lt;h3&gt;
  
  
  GitHub Copilot — 6/10
&lt;/h3&gt;

&lt;p&gt;Copilot added type hints quickly and correctly. It caught several obvious bugs. But it struggled with the big-picture refactor — understanding &lt;em&gt;why&lt;/em&gt; certain choices were made. The type hints were correct but superficial.&lt;/p&gt;

&lt;p&gt;When I asked it to "refactor this into smaller functions," it gave a reasonable split but broke the data pipeline ordering. I had to manually fix the function call chain.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Best for&lt;/strong&gt;: Tab completion and boilerplate. Not architecture.&lt;/p&gt;

&lt;h3&gt;
  
  
  Cursor — 7/10
&lt;/h3&gt;

&lt;p&gt;Cursor did better at understanding full file context. Its inline suggestions for type hints were solid. When I selected a 100-line block and asked "refactor this," it proposed a clean extraction with proper error handling.&lt;/p&gt;

&lt;p&gt;It missed the silent data loss bug on its own, but when I pointed at the specific region, it correctly identified the off-by-one error in the batch processing loop.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Best for&lt;/strong&gt;: Refactoring with context awareness.&lt;/p&gt;

&lt;h3&gt;
  
  
  Claude Code — 5/10
&lt;/h3&gt;

&lt;p&gt;Claude Code was too aggressive. When I said "refactor this file," it rewrote the entire thing from scratch — new structure, new patterns, everything. The result was cleaner code, but it changed behavior in subtle ways that would break production.&lt;/p&gt;

&lt;p&gt;To be fair, when I said "no, just add types and fix bugs," it did exactly that. But I had to catch its first attempt. That is not autonomy.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Best for&lt;/strong&gt;: Greenfield projects where you want a fresh architecture.&lt;/p&gt;

&lt;h3&gt;
  
  
  Windsurf — 6/10
&lt;/h3&gt;

&lt;p&gt;Solid basics. Good type inference. Decent refactor suggestions. But it asked for confirmation on &lt;em&gt;every single change&lt;/em&gt;. After 47 confirmations, I wanted to throw my laptop out the window.&lt;/p&gt;

&lt;p&gt;In cascade mode, it got more autonomous but also more error-prone. It changed a &lt;code&gt;dict&lt;/code&gt; to a &lt;code&gt;defaultdict&lt;/code&gt; without asking, which subtly changed error behavior.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Best for&lt;/strong&gt;: People who want fine-grained control over every change.&lt;/p&gt;

&lt;h3&gt;
  
  
  Aider — 4/10
&lt;/h3&gt;

&lt;p&gt;Aider struggled with the 400-line file. It kept losing context — making a change, then suggesting another that contradicted the first one. The refactor it proposed was correct in isolation but broke imports across the codebase.&lt;/p&gt;

&lt;p&gt;I had to explicitly say "keep all imports unchanged" for it to produce safe output.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Best for&lt;/strong&gt;: Small, well-defined tasks with clear boundaries.&lt;/p&gt;

&lt;h3&gt;
  
  
  Winner: Cursor (7/10)
&lt;/h3&gt;

&lt;p&gt;Best balance of autonomy and accuracy for refactoring legacy code.&lt;/p&gt;




&lt;h2&gt;
  
  
  Task 2: Greenfield Feature
&lt;/h2&gt;

&lt;p&gt;Build a service that:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Connects to a WebSocket stream (crypto exchange ticker)&lt;/li&gt;
&lt;li&gt;Transforms raw events into normalized records&lt;/li&gt;
&lt;li&gt;Batches writes to PostgreSQL (1000 records or 5 seconds)&lt;/li&gt;
&lt;li&gt;Handles reconnection with exponential backoff&lt;/li&gt;
&lt;li&gt;Implements backpressure — stop reading if DB queue exceeds 10,000&lt;/li&gt;
&lt;li&gt;Supports schema evolution — new fields should not crash the pipeline&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;I gave each tool the exact same specification paragraph. No starter code.&lt;/p&gt;

&lt;h3&gt;
  
  
  GitHub Copilot — 4/10
&lt;/h3&gt;

&lt;p&gt;Copilot is a tab-completion engine, not an architect. It wrote reasonable code line by line but had no sense of overall design. The WebSocket client was fine. The PostgreSQL writer was fine. But the connection between them — the part that actually matters — was a mess. No backpressure. No graceful shutdown. Thread-safety issues everywhere.&lt;/p&gt;

&lt;p&gt;I had to design the architecture myself and use Copilot as a faster typing tool.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Verdict&lt;/strong&gt;: Good pair programmer, bad system architect.&lt;/p&gt;

&lt;h3&gt;
  
  
  Cursor — 8/10
&lt;/h3&gt;

&lt;p&gt;Cursor impressed me here. It asked clarifying questions before writing code:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"Should the backpressure block the WebSocket reader or drop messages?"&lt;br&gt;
"Do you want schema evolution to be strict (reject unknown fields) or permissive (store them in a JSONB column)?"&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Then it generated a complete, working implementation. AsyncIO based. Proper connection management. Real backpressure with &lt;code&gt;asyncio.Queue(maxsize=10000)&lt;/code&gt;. Schema evolution via a JSONB overflow column. All in ~200 lines.&lt;/p&gt;

&lt;p&gt;I ran it against a test WebSocket server. It worked on the first try.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Verdict&lt;/strong&gt;: Closest thing to an "AI software engineer" I have seen.&lt;/p&gt;

&lt;h3&gt;
  
  
  Claude Code — 6/10
&lt;/h3&gt;

&lt;p&gt;Claude Code generated beautiful code. Clean architecture. Type annotations everywhere. Comprehensive error handling. It even added a health check endpoint and structured logging that I didn't ask for.&lt;/p&gt;

&lt;p&gt;The problem: it used &lt;code&gt;asyncio.gather()&lt;/code&gt; without proper error propagation. When the WebSocket connection dropped, the entire process silently hung instead of crashing. I caught it during testing, but this is the kind of bug that makes it to production if you trust the output without reading it.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Verdict&lt;/strong&gt;: Beautiful code, subtle bugs. Always review before shipping.&lt;/p&gt;

&lt;h3&gt;
  
  
  Windsurf — 7/10
&lt;/h3&gt;

&lt;p&gt;Strong implementation. Good structure. It chose a multi-process approach instead of async, which I disagreed with but it worked. The backpressure implementation was creative — semaphore-based throttle instead of queue size checking.&lt;/p&gt;

&lt;p&gt;It missed the schema evolution requirement entirely. I had to explicitly ask for it.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Verdict&lt;/strong&gt;: Solid but not thorough. You need to be the project manager.&lt;/p&gt;

&lt;h3&gt;
  
  
  Aider — 5/10
&lt;/h3&gt;

&lt;p&gt;Aider produced working code after three iterations. The first attempt had no error handling. The second added error handling but broke the batch writer. The third was functional but had a subtle race condition in the backpressure logic.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Verdict&lt;/strong&gt;: Feels like pair programming with a junior dev who reads docs fast.&lt;/p&gt;

&lt;h3&gt;
  
  
  Winner: Cursor (8/10)
&lt;/h3&gt;

&lt;p&gt;Its ability to ask clarifying questions and design a system end-to-end is unmatched.&lt;/p&gt;




&lt;h2&gt;
  
  
  Task 3: Debug Mystery
&lt;/h2&gt;

&lt;p&gt;A Node.js Express service in Kubernetes. Under load (&amp;gt;500 concurrent requests), ~3% return 502 Bad Gateway. No stack traces. No error logs. Health check works fine. Memory and CPU look normal.&lt;/p&gt;

&lt;p&gt;This bug had been open for 2 weeks. Two senior engineers had looked at it.&lt;/p&gt;

&lt;h3&gt;
  
  
  GitHub Copilot — 3/10
&lt;/h3&gt;

&lt;p&gt;Copilot is not a debugger. It suggested checking error handlers, adding logging, and looking at the reverse proxy config — the same things the human engineers already tried.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Verdict&lt;/strong&gt;: Useless for debugging.&lt;/p&gt;

&lt;h3&gt;
  
  
  Cursor — 6/10
&lt;/h3&gt;

&lt;p&gt;When I gave Cursor access to the full codebase, it noticed something: the service uses a connection pool for downstream HTTP calls, and the pool has a default timeout of 30s. But the service has middleware that sets a request timeout of 29s.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The bug&lt;/strong&gt;: When traffic spikes, connections queue up. Some requests hit the middleware timeout before the connection pool returns a connection. The middleware catches the timeout and returns a 502 — but the error happens outside the try-catch that logs errors. No log, no trace, just a 502.&lt;/p&gt;

&lt;p&gt;This was the actual bug. A cursed interaction between two timeouts implemented by two different people 6 months apart.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Verdict&lt;/strong&gt;: Actually useful for debugging. Reads the whole codebase.&lt;/p&gt;

&lt;h3&gt;
  
  
  Claude Code — 7/10
&lt;/h3&gt;

&lt;p&gt;Claude Code found the same bug as Cursor but faster. It read the middleware chain, the connection pool config, and the error handling in sequence and said:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"There is a 1-second gap between the middleware timeout (29s) and the pool timeout (30s). During this gap, requests are cancelled by the middleware but the error handler does not catch cancellation errors. Try adding &lt;code&gt;process.on('unhandledRejection', ...)&lt;/code&gt; and check if cancellation errors are being swallowed."&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;It was right. The fix was a 3-line change to the error handler.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Verdict&lt;/strong&gt;: Best debugger of the bunch. Reads code like a senior engineer.&lt;/p&gt;

&lt;h3&gt;
  
  
  Windsurf — 5/10
&lt;/h3&gt;

&lt;p&gt;Windsurf found the timeout mismatch but didn't connect it to the missing error logging. It said "these timeouts look close, maybe that's the problem?" — but didn't explain &lt;em&gt;why&lt;/em&gt; or &lt;em&gt;how&lt;/em&gt; to verify it.&lt;/p&gt;

&lt;p&gt;I had to do the actual debugging myself.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Verdict&lt;/strong&gt;: Hints at the answer but doesn't get you all the way there.&lt;/p&gt;

&lt;h3&gt;
  
  
  Aider — 4/10
&lt;/h3&gt;

&lt;p&gt;Aider couldn't handle this task. It doesn't have a "read the whole codebase and form a hypothesis" mode. It works on files you explicitly show it. By the time I had shown it all the relevant files, I had basically debugged it myself.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Verdict&lt;/strong&gt;: Not built for debugging.&lt;/p&gt;

&lt;h3&gt;
  
  
  Winner: Claude Code (7/10)
&lt;/h3&gt;

&lt;p&gt;Fastest to the correct diagnosis. Best at reasoning about system-level interactions.&lt;/p&gt;




&lt;h2&gt;
  
  
  Overall Scores
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Tool&lt;/th&gt;
&lt;th&gt;Task 1 (Refactor)&lt;/th&gt;
&lt;th&gt;Task 2 (Greenfield)&lt;/th&gt;
&lt;th&gt;Task 3 (Debug)&lt;/th&gt;
&lt;th&gt;&lt;strong&gt;Average&lt;/strong&gt;&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Cursor&lt;/td&gt;
&lt;td&gt;7&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;8&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;6&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;7.0&lt;/strong&gt;&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Claude Code&lt;/td&gt;
&lt;td&gt;5&lt;/td&gt;
&lt;td&gt;6&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;7&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;6.0&lt;/strong&gt;&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Windsurf&lt;/td&gt;
&lt;td&gt;6&lt;/td&gt;
&lt;td&gt;7&lt;/td&gt;
&lt;td&gt;5&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;6.0&lt;/strong&gt;&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;GitHub Copilot&lt;/td&gt;
&lt;td&gt;6&lt;/td&gt;
&lt;td&gt;4&lt;/td&gt;
&lt;td&gt;3&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;4.3&lt;/strong&gt;&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Aider&lt;/td&gt;
&lt;td&gt;4&lt;/td&gt;
&lt;td&gt;5&lt;/td&gt;
&lt;td&gt;4&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;4.3&lt;/strong&gt;&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;




&lt;h2&gt;
  
  
  The Verdict
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;If you write code for a living, get Cursor.&lt;/strong&gt; It is the only tool that consistently helps across all three task types. The $20/month is cheaper than one hour of your time.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;If you do a lot of debugging, add Claude Code to your toolkit.&lt;/strong&gt; It reasons about code differently — more like a senior engineer than a code completion engine.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;GitHub Copilot is fine if you already have it&lt;/strong&gt;, but it is not worth $10/month if you don't. It is a fancy autocomplete, not a coding assistant.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Aider is the best free option&lt;/strong&gt; if you are comfortable on the command line and don't mind hand-holding the AI.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Windsurf is Cursor but more annoying to use.&lt;/strong&gt; Skip it unless you really want that cascade feature.&lt;/p&gt;




&lt;h2&gt;
  
  
  The One Thing Nobody Tells You
&lt;/h2&gt;

&lt;p&gt;All of these tools make you faster at writing code. None of them make you faster at &lt;em&gt;thinking about what to build&lt;/em&gt;.&lt;/p&gt;

&lt;p&gt;If you don't know what you're doing, AI will help you build the wrong thing faster.&lt;/p&gt;

&lt;p&gt;The developers who get the most out of these tools are not the ones who prompt the best. They are the ones who already know what good code looks like and use AI to get there faster.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;AI is a force multiplier, not a replacement for judgment.&lt;/strong&gt;&lt;/p&gt;




&lt;p&gt;&lt;em&gt;If you found this useful, follow me on Dev.to for more honest tool reviews and practical engineering advice. No affiliate links, no sponsorship, just real testing.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>programming</category>
      <category>productivity</category>
      <category>career</category>
    </item>
  </channel>
</rss>
