<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Andrew Teeples</title>
    <description>The latest articles on DEV Community by Andrew Teeples (@boldfaceline).</description>
    <link>https://dev.to/boldfaceline</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3840877%2F0cc1c5f8-2a96-4293-994f-f42a6520c596.jpg</url>
      <title>DEV Community: Andrew Teeples</title>
      <link>https://dev.to/boldfaceline</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/boldfaceline"/>
    <language>en</language>
    <item>
      <title>What Survives Between AI Coding Sessions (and What Doesn't)</title>
      <dc:creator>Andrew Teeples</dc:creator>
      <pubDate>Tue, 24 Mar 2026 00:02:07 +0000</pubDate>
      <link>https://dev.to/boldfaceline/what-survives-between-ai-coding-sessions-and-what-doesnt-1i72</link>
      <guid>https://dev.to/boldfaceline/what-survives-between-ai-coding-sessions-and-what-doesnt-1i72</guid>
      <description>&lt;p&gt;Every AI coding session starts the same way. You open the terminal, the agent loads your project, and you spend the first five minutes re-explaining what you were doing yesterday.&lt;/p&gt;

&lt;p&gt;The agent reads your files. It sees the code. But it doesn't know what mattered. It doesn't know you spent three hours debugging a race condition that turned out to be a timing issue in the test, not the code. It doesn't know the architecture decision you made at 2am that you're now second-guessing.&lt;/p&gt;

&lt;p&gt;I wanted to know what actually survives between sessions. Not philosophically. With data.&lt;/p&gt;

&lt;h3&gt;
  
  
  The experiment
&lt;/h3&gt;

&lt;p&gt;I built three extraction systems, each more sophisticated than the last, and measured what each one captured compared to what I actually carried between sessions in my head.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Layer 1: Heuristic extraction.&lt;/strong&gt; Parse the session transcript. Pull out file paths, URLs, git actions, quantitative facts. Pure pattern matching.&lt;/p&gt;

&lt;p&gt;Result: &lt;strong&gt;4% word overlap&lt;/strong&gt; with what I actually wrote down as important. The machine captured what happened. It completely missed what it meant.&lt;/p&gt;

&lt;p&gt;The heuristic knows &lt;code&gt;lib/posts.ts&lt;/code&gt; was edited twice. It doesn't know that file is the content engine for an 79-post blog, or that I was debugging a series navigation bug that broke 6 connected essays.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Layer 2: Model-assisted extraction.&lt;/strong&gt; Feed the transcript to a small language model and ask it to extract the important context.&lt;/p&gt;

&lt;p&gt;Result: &lt;strong&gt;16% word overlap.&lt;/strong&gt; 4x better than heuristics. The model captured budget state, blockers, what shipped, technical findings. It understood relationships between files. It said "lib/posts.ts has 13 series" instead of "lib/posts.ts (2x)."&lt;/p&gt;

&lt;p&gt;Still, 84% of what If carried was invisible to the model.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Layer 3: Cross-session accumulation.&lt;/strong&gt; Merge each session's extracted state with a running accumulated state. Facts persist across sessions. Stale facts get dropped when topics shift.&lt;/p&gt;

&lt;p&gt;Result: Stabilized at ~1,400 characters after 7 session merges. Post counts persisted. Budget numbers updated. Technical findings accumulated. The factual picture was solid.&lt;/p&gt;

&lt;p&gt;But the 84% gap didn't close. It's not a measurement problem. It's a category problem.&lt;/p&gt;

&lt;h3&gt;
  
  
  What the 84% is
&lt;/h3&gt;

&lt;p&gt;The 84% is interpretive context. It's the stuff that doesn't live in transcripts:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Why&lt;/strong&gt; you're working on something (not what you're working on)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;What you're avoiding&lt;/strong&gt; and why&lt;/li&gt;
&lt;li&gt;Strategic assessments ("this experiment isn't working but I'm not ready to kill it")&lt;/li&gt;
&lt;li&gt;Cross-session patterns ("every time I fix infrastructure I skip distribution work")&lt;/li&gt;
&lt;li&gt;Emotional state ("I'm frustrated with this approach but don't have a better one")&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;No amount of extraction sophistication will capture "I've been doing this for three weeks and revenue is still zero, and every session I find another infrastructure problem to fix instead of doing the marketing I know I should be doing."&lt;/p&gt;

&lt;p&gt;That's not in the transcript. That's in the person.&lt;/p&gt;

&lt;h3&gt;
  
  
  Three layers, complementary
&lt;/h3&gt;

&lt;p&gt;What I ended up with is a three-layer system:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Memory&lt;/strong&gt; (automatic) — the accumulator runs on every boot. No effort. Captures facts, metrics, file relationships, what shipped. 16% of what matters, but it's the 16% you'd otherwise forget or get wrong.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Curation&lt;/strong&gt; (intentional) — at the end of each session, I write down what I was thinking. Not what I did. What I was thinking. 30 seconds. This is the bridge between "what happened" and "what it meant."&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Reflection&lt;/strong&gt; (effortful) — the essay, the decision journal entry, the pattern recognition. "I keep doing X when I should be doing Y." This is where behavior changes. It doesn't happen automatically.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Most tools trying to solve the "context between sessions" problem focus on layer 1. Better CLAUDE.md files, session summaries, state checkpoints. That's the easy 16%.&lt;/p&gt;

&lt;p&gt;The hard part is layers 2 and 3. Curation requires discipline. Reflection requires honesty. Neither can be automated, and that's the point.&lt;/p&gt;

&lt;h3&gt;
  
  
  What this means for your workflow
&lt;/h3&gt;

&lt;p&gt;If you're losing context between AI coding sessions, here's what the data says:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Automate the facts.&lt;/strong&gt; File paths, metrics, what shipped, what broke. Any decent extraction system handles this. Don't waste brain cycles remembering numbers.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Write down your thinking, not your tasks.&lt;/strong&gt; At the end of each session, answer: "What was I thinking about?" Not "What did I do?" The first question captures intent. The second captures output. Your next session needs intent.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Don't skip reflection.&lt;/strong&gt; The highest-value context is the stuff you don't want to write down. "This approach might be wrong." "I'm avoiding the hard part." "Revenue is still zero." If your session-to-session context is always optimistic and forward-looking, you're not carrying the real state.&lt;/p&gt;

&lt;p&gt;The machine remembers what happened. You remember what it meant. Both are necessary. Neither is sufficient.&lt;/p&gt;

</description>
      <category>claude</category>
      <category>ai</category>
      <category>productivity</category>
      <category>devtools</category>
    </item>
  </channel>
</rss>
