This is a submission for the Google AI Agents Writing Challenge: Learning Reflections
Few days ago, I thought AI agents were just chatbots with extra steps. Today, I have a working multi-agent research system deployed on GitHub. This is the story of my journey through Google and Kaggle's AI Agents Intensive Course—the challenges, breakthroughs, and everything I learned along the way.
The Starting Point: What I Thought I Knew
Before the course, my understanding of AI agents was surface-level at best. I knew about ChatGPT and basic prompting, but the concept of agents that could autonomously plan, use tools, and collaborate seemed almost magical. I had more questions than answers:
How do agents decide what to do next?
What makes an agent different from a simple API call?
How do multiple agents work together without chaos?
The Hands-On Experience: Building a Research Agent
The course's hands-on approach threw me into the deep end—and that's exactly what I needed. My main project was building a research agent system capable of:
Conducting web searches autonomously
Fetching and analyzing content from multiple sources
Synthesizing information into coherent reports
Making decisions about when to search deeper vs. when to conclude
The Technical Journey
Setting Up the Foundation:
Working with the repository structure taught me that AI agents aren't monolithic. They're composed of:
Tools (web search, content fetching, data processing)
Planning mechanisms (deciding what to do next)
Memory systems (maintaining context across operations)
Orchestration layers (coordinating multiple agents)
The Git Learning Curve:
Beyond AI, I faced practical challenges like managing code with Git. Merge conflicts, branch management, and pushing updates to GitHub became second nature through trial and error. Every git push that succeeded felt like a small victory.
Conceptual Breakthroughs
- Agents Are About Autonomy, Not Intelligence The biggest mindset shift: an AI agent isn't necessarily "smarter" than a language model—it's more autonomous. The key is giving the model:
The ability to use tools
A framework for deciding when to use them
A way to evaluate if it's making progress
This clicked when I saw my research agent decide on its own to do follow-up searches based on initial findings.
- Tool Use Is the Superpower Before: I thought the language model's knowledge was the limitation. After: I realized the real power is in tool use. My agent could:
Search the live web (breaking past the knowledge cutoff)
Fetch entire web pages (getting full context)
Process structured data (extracting insights)
The LLM became the "brain" that orchestrated these tools, not the sole source of information.
- Iteration Over Perfection Early on, I wanted my agent to be perfect on the first try. But I learned that agent development is iterative:
Start with simple tool calls
Test with real queries
Observe where it fails
Add guardrails and improvements
Repeat
My research agent went through multiple versions, each one slightly better at knowing when to search deeper vs. when it had enough information.
Challenges I Overcame
Challenge 1: The Model Wasn't Following Instructions
The Problem: My agent would sometimes ignore available tools and try to answer from memory alone.
The Solution: I learned about prompt engineering for agents:
Being explicit about available tools
Providing examples of tool use
Structuring the system prompt to encourage tool adoption
Adding instructions about when each tool is appropriate
Challenge 2: Context Window Management
The Problem: After multiple searches, the conversation history would balloon, causing errors.
The Solution: Understanding that agents need memory management:
Summarizing previous findings
Keeping only relevant context
Designing the agent to work within token limits
This taught me that constraints breed creativity—working within limits made my agent more efficient.
Challenge 3: Debugging Black Box Decisions
The Problem: When my agent made a bad decision, I couldn't figure out why.
The Solution: Adding logging and intermediate outputs:
Printing the agent's reasoning
Tracking which tools were called
Observing the decision-making process
Debugging AI agents is different from debugging traditional code—you're debugging reasoning, not just logic.
Challenge 4: Version Control Chaos
The Problem: Git merge conflicts when trying to update my repository.
The Solution: Learning proper Git workflow:
Always pull before pushing
Understanding merge vs. rebase
Using branches for experimentation
Reading error messages carefully (they actually help!)
How My Understanding Evolved
Week 1 Thinking vs. Week 2 Reality
Before the course:
"AI agents are just chatbots with APIs"
"The model needs to know everything"
"More parameters = better agent"
After the course:
AI agents are autonomous systems with reasoning loops
The model orchestrates tools—tools provide knowledge
Good design > raw model size
The Paradigm Shift
The course shifted my thinking from:
"What can the AI know?" → "What can the AI do?"
"How do I prompt this?" → "How do I architect this?"
"Single interaction" → "Multi-step reasoning"
Key Takeaways
- Start Simple, Then Scale My best progress came when I:
Built one tool at a time
Tested thoroughly before adding complexity
Celebrated small wins
- Agents Need Guardrails Without limits, agents can:
Loop infinitely
Make unnecessary tool calls
Exceed context windows
Good agent design includes knowing when to stop.
- The Human Element Matters Even with autonomous agents, human oversight is crucial:
Validating outputs
Guiding overall direction
Deciding what tools to provide
We're building assistants, not replacements.
- Documentation Is Your Friend Reading the API docs, understanding tool parameters, and following best practices saved me hours of debugging. When in doubt, RTFM (Read The Fine Manual). What's Next This course was just the beginning. I'm now equipped to:
Build more specialized agents (I'm thinking about a code review agent next)
Experiment with agent collaboration (multiple agents working together)
Deploy agents in real-world scenarios
Continue learning about LangChain, LangGraph, and other frameworks
The research agent project is live on my GitHub, and I'm actively improving it based on what I learned.
Final Thoughts
Five days feels like both forever and no time at all. I went from theoretical knowledge to practical application, from confusion to clarity. The hands-on labs were challenging but invaluable—there's no substitute for actually building something.
To anyone considering this course: embrace the confusion. The moments where you're stuck are where the learning happens. Every error message, every failed deployment, every Git conflict is a lesson in disguise.
Thank you to Google and Kaggle for this intensive experience. I'm leaving with not just knowledge, but a working project and the confidence to build more.
GitHub Repository: github.com/Dessyonpoint/research-agent
What I'm building next: A mask detection model integrated with an AI agent for real-time analysis (combining my ongoing computer vision project with agent capabilities)
Connect with me: Feel free to check out my projects and let's build the future of AI together!
Top comments (0)