<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Danda Pranav Krishna</title>
    <description>The latest articles on DEV Community by Danda Pranav Krishna (@21bq1a4232).</description>
    <link>https://dev.to/21bq1a4232</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3476900%2Fdc686ba5-b459-43ae-af61-2f37f4661779.jpeg</url>
      <title>DEV Community: Danda Pranav Krishna</title>
      <link>https://dev.to/21bq1a4232</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/21bq1a4232"/>
    <language>en</language>
    <item>
      <title>Memora — Adaptive AI Agents with Memory (No Fine-Tuning Required)</title>
      <dc:creator>Danda Pranav Krishna</dc:creator>
      <pubDate>Wed, 03 Sep 2025 07:35:05 +0000</pubDate>
      <link>https://dev.to/21bq1a4232/memora-adaptive-ai-agents-with-memory-no-fine-tuning-required-2g44</link>
      <guid>https://dev.to/21bq1a4232/memora-adaptive-ai-agents-with-memory-no-fine-tuning-required-2g44</guid>
      <description>&lt;p&gt;The Problem with Today’s AI Agents&lt;/p&gt;

&lt;p&gt;Large Language Models (LLMs) are powerful, but they don’t truly learn from experience.&lt;br&gt;
Each interaction is isolated — no memory of past attempts, no cumulative knowledge.&lt;/p&gt;

&lt;p&gt;Fine-tuning can help, but it’s:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Expensive&lt;/li&gt;
&lt;li&gt;Rigid&lt;/li&gt;
&lt;li&gt;Slow to iterate&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If we want truly adaptive agents, we need a better paradigm.&lt;/p&gt;

&lt;p&gt;The Idea: Memory-Augmented Reinforcement Learning&lt;/p&gt;

&lt;p&gt;Instead of retraining the model itself, Memora introduces memory into the loop.&lt;/p&gt;

&lt;p&gt;Episodic Memory — stores past experiences (success + failure).&lt;/p&gt;

&lt;p&gt;Case Retrieval — brings up the most relevant past examples for new tasks.&lt;/p&gt;

&lt;p&gt;Memory Rewriting — updates knowledge dynamically with feedback.&lt;/p&gt;

&lt;p&gt;This shifts the agent’s learning from parameter updates → to retrieval + reasoning.&lt;/p&gt;

&lt;p&gt;How Memora Works&lt;/p&gt;

&lt;p&gt;The architecture follows a Planner–Executor cycle:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Meta-Planner (System 2):&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Strategically breaks down complex problems.&lt;/p&gt;

&lt;p&gt;Leverages memory for analogical reasoning.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Executor (System 1):&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Executes steps sequentially.&lt;/p&gt;

&lt;p&gt;Writes results back to memory.&lt;/p&gt;

&lt;p&gt;This means the agent improves with experience  without touching the base model weights.&lt;/p&gt;

&lt;p&gt;Key Results&lt;/p&gt;

&lt;p&gt;GAIA benchmark: 87.88% validation (outperforming GPT-4 baseline).&lt;/p&gt;

&lt;p&gt;DeepResearcher benchmark: +4.7–9.6% gain on out-of-domain tasks.&lt;/p&gt;

&lt;p&gt;Local LLMs (Qwen2.5-14B, LLaMA): achieved near GPT-4 performance — on a consumer MacBook (M4).&lt;/p&gt;

&lt;p&gt;Why This Matters&lt;/p&gt;

&lt;p&gt;Continual learning without retraining.&lt;/p&gt;

&lt;p&gt;Cost efficiency — runs on everyday hardware.&lt;/p&gt;

&lt;p&gt;Interpretability — every decision can be traced back to memory.&lt;/p&gt;

&lt;p&gt;Scalability — agents adapt in real time.&lt;/p&gt;

&lt;p&gt;Try It Yourself&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;# Install Ollama
curl -fsSL https://ollama.ai/install.sh | sh

# Pull a local model
ollama pull qwen2.5:14b

# Clone the repo
git clone https://github.com/Agent-on-the-Fly/Memora
cd Memora &amp;amp;&amp;amp; pip install -r requirements.txt

# Run the agent
python client/agent.py
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;👉 &lt;a href="https://github.com/21bq1a4232/Memora" rel="noopener noreferrer"&gt;GitHub Repo&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Final Thoughts&lt;/p&gt;

&lt;p&gt;The future of AI isn’t just about building bigger models — it’s about building smarter agents with memory.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsischhg82polodyd4ju8.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsischhg82polodyd4ju8.png" alt=" " width="800" height="631"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Memora shows that experience &amp;gt; parameters.&lt;br&gt;
And this shift may redefine how we build and deploy intelligent systems.&lt;/p&gt;

&lt;p&gt;Cross-posted from my &lt;a href="https://agentfly.hashnode.dev/adaptive-llm-agents-that-learn-without-fine-tuning" rel="noopener noreferrer"&gt;Hashnode blog&lt;/a&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>softwareengineering</category>
      <category>webdev</category>
    </item>
  </channel>
</rss>
