DEV Community

Cover image for From User to Builder: My Honest Review of Google's 5-Day AI Agents Intensive Course with Kaggle
Mahesh Jagtap
Mahesh Jagtap

Posted on

From User to Builder: My Honest Review of Google's 5-Day AI Agents Intensive Course with Kaggle

This is a submission for the Google AI Agents Writing Challenge: Learning Reflections

From User to Builder

Over the past five days, I immersed myself in Google and Kaggle’s AI Agents Intensive, a hands-on learning sprint designed to help participants understand, build, and deploy AI agents using practical tools and real-world challenges. What began as curiosity quickly evolved into a structured, insightful journey into the future of intelligent automation.


🌱 Day 1 — Foundations of AI Agents

The program kicked off with the core concepts: What are AI agents? How do they perceive, reason, and act?
I explored agent architectures, from simple reactive designs to more advanced planning-based models. The highlight of the day was experimenting with pre-built agents on Kaggle and observing how they handled tasks autonomously. It was the first moment I realized how transformative agent-driven workflows can be.


🛠️ Day 2 — Tools, Frameworks & Notebook Walkthroughs

This day focused on the practical ecosystem behind AI agents.
I learned how to use Kaggle's notebook environment, integrated APIs, and Google’s developer tools to set up the scaffolding for agent experiments.
Hands-on exercises included:

  • interacting with agent toolsets
  • modifying simple agent behaviors
  • experimenting with prompt engineering for task optimization

It was my first real taste of building—not just learning.


🤖 Day 3 — Building My First Agent

This was the breakthrough moment.
I built a functional AI agent capable of performing a multi-step task on its own.
I learned how to:

  • define agent goals
  • provide tools and constraints
  • evaluate the agent’s reasoning trace
  • refine its behavior through iterative feedback

Seeing my agent complete tasks end-to-end felt incredibly rewarding.


🚀 Day 4 — Advanced Agent Workflows & Optimization

On Day 4, we went deeper.
The focus shifted to agent robustness: How do you make an agent reliable? Efficient? Safe?
I explored techniques such as:

  • chaining tools
  • adding memory and state
  • improving reasoning patterns
  • using evaluation benchmarks from Kaggle

This day challenged me to think like a system designer, not just a user.


🏁 Day 5 — Capstone Challenge & Reflection

The final day culminated in a mini-project: create an agent capable of solving a realistic problem with minimal intervention.
My agent wasn’t perfect, but it worked—and the process taught me more than success alone ever could.

This journey changed the way I think about AI:
It’s not just about models or prompts anymore. It’s about autonomous, goal-driven systems that can collaborate with humans to streamline tasks, explore data, and solve meaningful problems.


🌟 What I’m Taking Away

  • AI agents are the next major step in everyday AI applications.
  • Even beginners can build functional agents with the right tools.
  • Experimentation is the fastest way to understand how these systems think.
  • The future of work will be shaped by human–agent collaboration.

Multi-Agent Customer Support Assistant — Capstone Project Overview

This project implements a simple but fully functional Multi-Agent Customer Support Assistant built for the Enterprise Agents track.
The purpose of this system is to demonstrate how multiple specialized agents can work together to automate a real business workflow—in this case, handling customer messages in a support environment.
Although the agents are lightweight and rule-based, the architecture clearly represents how multi-agent frameworks operate in enterprise settings: through specialization, coordination, and automated decision-making.

What This System Does

When a user sends a message (like “I need a refund” or “My invoice amount is wrong”), the system processes it using three different agents, each responsible for a specific task:

1. Intent Agent (Understands the Customer’s Message)

This agent analyzes the message and identifies its intent (refund, cancellation, billing issue, etc.) and urgency level (low, medium, high).
Even with simple rules, this agent demonstrates classification, routing, and task identification—core elements of enterprise automation.

*2. Reply Agent (Generates a Professional Response)
*

Once the intent is identified, the Reply Agent produces a short, clean, professional customer support reply.
This simulates how enterprises use AI to draft emails, chat responses, and automated replies for customer tickets.

3. Escalation Agent (Decides When Human Support Is Needed)

Not every customer message can be solved automatically.
This agent checks urgency and intent, and determines whether the issue requires escalation to a human support agent.
It produces escalation notes and reasons—mirroring how real businesses prioritize and triage tickets.

4. Coordinator Agent (The “Brain” of the System)

The Coordinator receives the message, calls the three specialized agents, collects their outputs, and returns a complete response package containing:

The predicted intent

The urgency level

The auto-generated reply

The escalation decision

A clean JSON output

This shows how multi-agent systems rely on orchestration, not just isolated decision-making.

Why I Built This Project

For the Enterprise Agents track, Kaggle requires the demonstration of multi-agent collaboration applied to a business problem.
I chose customer support automation because:

1. It is a real and common enterprise workflow

Companies receive thousands of customer tickets every day.
Automating the first layer of classification and response can save businesses a lot of time.

2. Easy to understand and demonstrate

Agents in this notebook have clear responsibilities and predictable outputs.
Judges and users can easily see how each agent contributes to the final answer.

3. A perfect fit for multi-agent architecture

Customer support naturally splits into:

Understanding the message

Generating a reply

Making escalation decisions

This makes it ideal for demonstrating agent specialization.

4. Lightweight but practical

The project uses simple rule-based logic instead of heavy models, making it:

Fast to run

Easy to understand

Safe to execute without external API calls

But the structure is extensible—LLMs can replace each agent for more advanced versions.

5. Meets all Kaggle agent competition requirements

📺 Project Overview Video (2 minutes)

https://youtu.be/ZfRAe9AJVfU?si

🎬Capstone Project Writeup

https://kaggle.com/competitions/agents-intensive-capstone-project/writeups/new-writeup-1764584691566

A huge thank you to the Google and Kaggle teams for putting this together. I highly recommend these materials to any developer or architect serious about building the next generation of AI.

Top comments (0)