Your AI coding agent doesn't remember yesterday.
You spent an hour debugging a tricky race condition, the AI understood every nuance — and this morning it asks you to "explain the project architecture." Again.
By the end of this post, you'll have persistent memory working across sessions in under 30 seconds. I'll show you the exact terminal output at every step so you can follow along.
Prerequisites
- Python 3.9+
- Any AI coding agent (Copilot, Claude Code, Cursor, Trae)
- A project you're actively working on
Step 1: Install and Initialize (15 seconds)
pip install fcontext
Successfully installed fcontext-1.0.0
Navigate to your project and initialize:
cd your-project
fcontext init
✓ Created .fcontext/
✓ Generated _README.md
✓ Generated _workspace.map
That's it. You now have a .fcontext/ directory:
.fcontext/
├── _README.md # Project summary — AI reads this first
├── _workspace.map # Auto-generated project structure
├── _topics/ # Where AI saves session knowledge
├── _requirements/ # Optional: track stories/tasks/bugs
└── _cache/ # Optional: converted binary docs
Everything is plain Markdown. No database, no cloud, no API keys.
Step 2: Enable Your Agent (15 seconds)
fcontext enable copilot
✓ Generated .github/instructions/fcontext.instructions.md
✓ Copilot will now read .fcontext/ on every session
Using a different agent? Swap the name:
fcontext enable claude # → .claude/rules/fcontext.md
fcontext enable cursor # → .cursor/rules/fcontext.md
fcontext enable trae # → .trae/rules/fcontext.md
Each agent gets instructions in its native config format. No plugins, no extensions — fcontext generates the standard config files that your agent already knows how to read.
Want multiple agents? Enable them all:
fcontext enable copilot && fcontext enable claude && fcontext enable cursor
They all read the same .fcontext/ data. Switch agents freely — no context loss.
Step 3: Verify It Works
Check what you've set up:
fcontext status
fcontext status
Initialized: yes
Agents: copilot, claude
Topics: 0
Requirements: 0
Cached files: 0
Now open your AI agent and start coding normally. Here's what happens behind the scenes:
-
Session starts → Agent reads
_README.mdand_topics/ - You work → Agent builds understanding of your project
-
Important discovery → Agent saves it to
_topics/your-topic.md - Session ends → Knowledge persisted on disk
- Next session → Agent reads everything from step 1 — full continuity
Real Example: Before and After
Before fcontext
# Monday session
You: "Help me debug the auth token refresh"
AI: [2 hours of deep debugging, finds race condition, adds mutex]
# Tuesday session
You: "What about edge cases in the auth fix?"
AI: "Could you provide more context about what you're working on?"
# 😤 starts re-explaining from scratch
After fcontext
# Monday session — same debugging work
# AI saves to .fcontext/_topics/auth-token-debugging.md:
# - Race condition: concurrent API calls trigger duplicate refresh
# - Fix: added mutex + storage key dedup + debounce
# - TODO: test edge case with expired refresh token
# Tuesday session
You: "What about edge cases in the auth fix?"
AI: "Based on yesterday's debugging, the mutex handles concurrent
refreshes. The edge case to test is an expired refresh token
during an active request. Want me to write that test?"
No re-explaining. The AI picks up exactly where you left off.
Bonus: Index Your Documents
Got PDFs, DOCX, or XLSX files in your project? Your AI can't read binary files — but fcontext can convert them:
fcontext index docs/
✓ Indexed docs/product-requirements.pdf → _cache/docs/product-requirements.pdf.md
✓ Indexed docs/api-spec.docx → _cache/docs/api-spec.docx.md
2 files indexed
Now your AI can reference those documents directly. No more copy-pasting from PDFs.
Supported formats: PDF, DOCX, XLSX, PPTX, Keynote, EPUB.
Bonus: Track Requirements
If your project has user stories or tasks scattered across Slack and docs:
fcontext req add "OAuth login flow" -t story
fcontext req add "Support Google provider" -t task --parent STORY-001
fcontext req set TASK-001 status in-progress
fcontext req board
📋 Board
TODO IN-PROGRESS DONE
───────── ───────── ────
TASK-001
Support Google
provider
STORY-001
OAuth login
flow
Your AI reads _requirements/ and builds against tracked specs — not guesses.
Common Gotchas
"My AI didn't read .fcontext/ on first session"
After fcontext enable, tell your AI: "Read .fcontext/_README.md and update it with the project info." It needs one nudge, then it maintains the file automatically.
"Can I git-commit .fcontext/?"
Yes — and you should. Your teammates pull the repo and get the same context. Their AI instantly knows the project.
git add .fcontext/
git commit -m "add project context"
"What if I want to start fresh?"
fcontext reset
Gone. Clean slate.
The Numbers
I tracked my context re-explanation time for two weeks. Week 1 without fcontext, week 2 with:
| Metric | Without | With fcontext |
|---|---|---|
| Daily context setup time | ~12 min | ~0 min |
| Agent switching overhead | ~10 min | 0 min |
| Weekly total waste | ~60 min | ~3 min |
The time saving is nice. But the real win is answer quality — an AI with accumulated project context gives better, more consistent responses than one starting from zero every morning.
TL;DR
pip install fcontext
fcontext init
fcontext enable copilot # or: claude, cursor, trae
30 seconds. Your AI now remembers.
GitHub: github.com/lijma/agent-skill-fcontext
What's your current workaround for AI context loss? Curious how others are dealing with this — drop a comment 👇
Top comments (0)