<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: FigCanvas</title>
    <description>The latest articles on DEV Community by FigCanvas (@figcanvas).</description>
    <link>https://dev.to/figcanvas</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3928140%2F5f9364c0-b3f8-49f6-b17d-d3c49ebd3814.png</url>
      <title>DEV Community: FigCanvas</title>
      <link>https://dev.to/figcanvas</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/figcanvas"/>
    <language>en</language>
    <item>
      <title>Why text-to-image AI keeps failing at scientific figures (and what actually works)</title>
      <dc:creator>FigCanvas</dc:creator>
      <pubDate>Wed, 13 May 2026 00:12:02 +0000</pubDate>
      <link>https://dev.to/figcanvas/why-text-to-image-ai-keeps-failing-at-scientific-figures-and-what-actually-works-2m9e</link>
      <guid>https://dev.to/figcanvas/why-text-to-image-ai-keeps-failing-at-scientific-figures-and-what-actually-works-2m9e</guid>
      <description>&lt;p&gt;Over the past month I've been trying to replace my "spend half a day in Illustrator drafting figures" workflow with AI tools. I tried Midjourney, GPT-Image-1, DALL-E, Stable Diffusion fine-tunes, generic diagram-from-text tools, and finally a purpose-built &lt;a href="https://figcanvas.com/" rel="noopener noreferrer"&gt;scientific illustration tool&lt;/a&gt;. Only the last one actually worked for figures destined for a peer-reviewed paper. This post is the autopsy on why.&lt;/p&gt;

&lt;h2&gt;
  
  
  The three failure modes of generic image AI on figures
&lt;/h2&gt;

&lt;h3&gt;
  
  
  1. Label hallucinations
&lt;/h3&gt;

&lt;p&gt;You ask for a figure showing "PCR cycling steps: denaturation, annealing, extension," and the model writes "Denaturition", "Aneling", and "Estention" inside the boxes. Or it gets the words right but spells "DNA" as a five-letter blob. This happens because image-gen models treat text as pixels; they don't know it's text.&lt;/p&gt;

&lt;p&gt;Workarounds people try:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Generate the image, then edit text in Photoshop. Works, but it's manual and removes the speed advantage.&lt;/li&gt;
&lt;li&gt;Use models with stronger text rendering (Flux 1.1 Pro, Ideogram). Better, but still wrong ~20% of the time, and you don't see which 20% until you've already exported.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;For a journal figure, the failure mode is invisible until a reviewer screenshots a mislabeled box and tells you to redo Figure 3.&lt;/p&gt;

&lt;h3&gt;
  
  
  2. Layouts don't iterate
&lt;/h3&gt;

&lt;p&gt;This is the real killer. Say the model gives you a four-panel figure: A, B, C, D. Reviewer asks: "Add a fifth panel showing the control condition."&lt;/p&gt;

&lt;p&gt;In pixels, there is no "add a panel." The only way to edit is to re-prompt. The new image will not preserve the exact layout, colors, fonts, or sizing of panels A-D. Every revision starts from scratch.&lt;/p&gt;

&lt;p&gt;Real-world cost: in a recent paper I had three revisions over six months. With pixel-AI tools, that's three from-scratch redraws. With Illustrator, it's three quick edits. With a structured-canvas tool that holds the figure as boxes/arrows/labels under the hood, it's three "add panel E" instructions, no redrawing.&lt;/p&gt;

&lt;h3&gt;
  
  
  3. Wrong visual vocabulary
&lt;/h3&gt;

&lt;p&gt;General image AIs are trained on stock photos, art, memes — not on scientific publications. The "diagram" they produce is the cartoony kind you'd see on a tech blog: 3D glossy boxes, comic-style arrows, gradient fills. Journals expect 2D, line-weight-controlled, color-restrained, vector outputs.&lt;/p&gt;

&lt;p&gt;You can prompt your way around some of this ("flat, minimal, journal style"), but the visual primitives the model knows are still pop-art primitives. It doesn't know what a cell membrane looks like in a methods schematic, or what a Sankey diagram for ecological flow should look like.&lt;/p&gt;

&lt;h2&gt;
  
  
  What actually works: a structured canvas with science-aware primitives
&lt;/h2&gt;

&lt;p&gt;The pattern that breaks all three problems: keep the figure as a &lt;strong&gt;structured representation&lt;/strong&gt; (boxes, arrows, labels, panels) underneath the natural-language prompt, and only render to pixels at export time.&lt;/p&gt;

&lt;p&gt;That way:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Text is text. It can't hallucinate spelling.&lt;/li&gt;
&lt;li&gt;"Add panel E" is a real operation on the structure.&lt;/li&gt;
&lt;li&gt;The library of primitives can be science-shaped: receptor cartoons, organelles, pathway arrows, multi-panel grids with consistent spacing.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The tool I landed on after the experiment is &lt;a href="https://figcanvas.com/" rel="noopener noreferrer"&gt;figcanvas.com&lt;/a&gt; — paste a Methods paragraph, get a structured first draft, iterate per panel with plain English, export vector. The first-draft quality isn't perfect (it sometimes drops a label when reshuffling), but the iteration loop is the win: I went from "empty Illustrator file" to a clean methods schematic in 25 minutes that used to be a half-day job. More importantly, when the second reviewer asks for changes, those changes take 10 minutes instead of two hours.&lt;/p&gt;

&lt;h2&gt;
  
  
  TL;DR
&lt;/h2&gt;

&lt;p&gt;Image-gen AI is great for thumbnails, blog hero images, and concept art. For scientific figures it's a trap because:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Text inside images is unreliable.&lt;/li&gt;
&lt;li&gt;Edits aren't compositional.&lt;/li&gt;
&lt;li&gt;The visual style is wrong for journals.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Pick a tool that treats the figure as structure, not pixels.&lt;/p&gt;




&lt;p&gt;If you write papers and resent every minute spent in Illustrator, the structured-canvas approach is worth a weekend trial. Even if you don't switch tools long-term, you'll learn what the AI tooling actually can and can't do for academic work in 2026.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>science</category>
      <category>tutorial</category>
      <category>productivity</category>
    </item>
  </channel>
</rss>
