Introduction:
Meetings are powerful — but remembering what was discussed is often a challenge. Important decisions, task assignments, and responsibilities often get lost in chats or notes.
To solve this, I built an AI Meeting Memory Agent that can:
Store meeting notes
Remember past conversations
Answer questions using memory
Use a fast LLM (Groq) for intelligent responses
This project demonstrates how AI agents with persistent memory can improve real-world productivity.
Problem statement:
In most teams:
Meeting notes are scattered
People repeatedly ask the same questions
Tasks get forgotten
Context is lost over time
We needed a system that doesn’t just chat — but remembers.
Solution:
I built an AI agent that combines:
🔹 Persistent Memory
Stores all meeting notes in a local JSON file.
🔹 Smart Retrieval
When a user asks a question, the system retrieves relevant past notes.
🔹 AI Reasoning (Groq LLM)
Uses Groq’s ultra-fast LLM to generate intelligent answers based on memory.
🔹 Streamlit UI
A simple, interactive web interface for users.
Tech Stack:
Python
Streamlit
Grouq API
JSON
dotenv
System Architecture:
User enters meeting note or question
System checks if input is a question
If note → store in memory
If question → fetch memory + send to Groq
Groq generates contextual answer
Response is displayed in UI
Key Feature — Memory + Intelligence
Unlike traditional chatbots, this system:
✔ Remembers past inputs
✔ Uses context to answer questions
✔ Improves over time
✔ Acts like a real AI assistant
Example:
Input:
Shiv will attend the meeting
Sri is doing backend
Question:
Who is doing backend?
Output:
Sri is doing backend
Why Groq?
I used Groq because:
Extremely fast inference
Free tier available
Supports powerful open-source models
Ideal for real-time AI agents
UI Preview
The interface includes:
Input box for notes/questions
Submit button
Memory viewer
Clear memory option
Challenges Faced:
API key integration issues
Model deprecation errors
Indentation bugs in Python
Memory retrieval tuning
Conclusion:
This project demonstrates how AI agents with memory can transform simple chat systems into intelligent assistants.
By combining:
Memory systems
LLM reasoning
Fast inference (Groq)
We can build practical tools for real-world productivity.
Top comments (0)