<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: PracticalAIGuy</title>
    <description>The latest articles on DEV Community by PracticalAIGuy (@practicalaiguy_ba30448492).</description>
    <link>https://dev.to/practicalaiguy_ba30448492</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3709943%2F3db1bf0b-cda4-4f67-8075-020a2be72567.jpeg</url>
      <title>DEV Community: PracticalAIGuy</title>
      <link>https://dev.to/practicalaiguy_ba30448492</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/practicalaiguy_ba30448492"/>
    <language>en</language>
    <item>
      <title>Visualizing the "Black Box": I traced a single prompt through an AI engine to see how it actually works</title>
      <dc:creator>PracticalAIGuy</dc:creator>
      <pubDate>Sun, 25 Jan 2026 20:02:29 +0000</pubDate>
      <link>https://dev.to/practicalaiguy_ba30448492/visualizing-the-black-box-i-traced-a-single-prompt-through-an-ai-engine-to-see-how-it-actually-1l4l</link>
      <guid>https://dev.to/practicalaiguy_ba30448492/visualizing-the-black-box-i-traced-a-single-prompt-through-an-ai-engine-to-see-how-it-actually-1l4l</guid>
      <description>&lt;p&gt;We talk constantly about the implications of AI—the ethics, the jobs, the singularity. But very few of us have a concrete mental model of the engineering that makes it happen.&lt;/p&gt;

&lt;p&gt;We know it's not magic. We know it's math. But what does that math look like in action?&lt;/p&gt;

&lt;p&gt;I wanted to bridge the gap between high-level hype and low-level code. So, I built a visual breakdown that &lt;strong&gt;traces the life of a single prompt: "Write a poem about a robot.&lt;/strong&gt;"&lt;/p&gt;

&lt;p&gt;I followed that prompt through the entire neural pipeline—Tokenization, Embeddings, Attention, and the KV Cache—to visualize exactly how a machine "thinks."&lt;/p&gt;

&lt;p&gt;Here is the full 16-minute visual breakdown:&lt;/p&gt;

&lt;p&gt;

  &lt;iframe src="https://www.youtube.com/embed/x-XkExN6BkI"&gt;
  &lt;/iframe&gt;


&lt;/p&gt;

&lt;p&gt;🧠 &lt;strong&gt;The Core Concepts (Visualized)&lt;/strong&gt;&lt;br&gt;
If you don't have time for the video right now, here are the top three mechanical analogies I use to replace the abstract jargon.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Embeddings are a "Grocery Store"&lt;/strong&gt;
How does a computer understand that "King" - "Man" + "Woman" = "Queen"? It’s not looking up definitions. It’s looking up locations.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Imagine a massive, hyper-dimensional Grocery Store.&lt;/p&gt;

&lt;p&gt;Apples are shelved next to Bananas.&lt;/p&gt;

&lt;p&gt;"King" is in the "Royalty" aisle.&lt;/p&gt;

&lt;p&gt;"Robot" is in the "Technology" aisle.&lt;/p&gt;

&lt;p&gt;When the AI sees a word, it doesn't read it; it turns it into coordinates (vectors). This allows it to understand relationships based on "distance" rather than definitions.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Attention is a "Cocktail Party"&lt;/strong&gt;
The biggest breakthrough in Generative AI was the Attention Mechanism. But how does it work?&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Think of it like a loud Cocktail Party. You act as the AI. You are surrounded by noise (tokens). Most of it is irrelevant. But if someone shouts your name, or a topic you care about, you snap to attention.&lt;/p&gt;

&lt;p&gt;The model does this mathematically. When it processes the word "Bank", it scans the rest of the sentence (the room).&lt;/p&gt;

&lt;p&gt;If it hears "River," it pays attention to the nature meaning of Bank.&lt;/p&gt;

&lt;p&gt;If it hears "Money," it pays attention to the financial meaning.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;The Context Window is a "Workbench"&lt;/strong&gt;
We often worry about AI "forgetting" things in long conversations. This isn't a memory failure; it's a space failure.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;The Context Window isn't an infinite brain; it’s a physical workbench. You can only fit 8,000 (or 128k) tools on the bench. Once it's full, if you add a new tool, the oldest one falls off the edge.&lt;/p&gt;

&lt;p&gt;Why this matters&lt;br&gt;
Understanding these mechanics removes the fear of the "Black Box." When you see that "Hallucinations" are just probabilistic guesses based on vector math, the AI feels less like a magic spirit and more like a powerful engine.&lt;/p&gt;

&lt;p&gt;I also cover RLHF (Reinforcement Learning from Human Feedback) in the video, explaining how we train the "Wild Wolf" (Base Model) into a "Helpful Dog" (Instruct Model).&lt;/p&gt;

&lt;p&gt;Let me know if these analogies help you grasp the science behind the hype! 👇&lt;/p&gt;

</description>
      <category>ai</category>
      <category>science</category>
      <category>education</category>
      <category>robotics</category>
    </item>
    <item>
      <title>I traced a single prompt through an LLM engine to see how it actually works (Visual Breakdown)</title>
      <dc:creator>PracticalAIGuy</dc:creator>
      <pubDate>Sun, 25 Jan 2026 15:13:31 +0000</pubDate>
      <link>https://dev.to/practicalaiguy_ba30448492/i-traced-a-single-prompt-through-an-llm-engine-to-see-how-it-actually-works-visual-breakdown-2kbj</link>
      <guid>https://dev.to/practicalaiguy_ba30448492/i-traced-a-single-prompt-through-an-llm-engine-to-see-how-it-actually-works-visual-breakdown-2kbj</guid>
      <description>&lt;p&gt;We all use the API. We send a JSON payload to /v1/chat/completions, wait a few hundred milliseconds, and get a magical response back.&lt;/p&gt;

&lt;p&gt;But as an engineer, the "Black Box" nature of AI bothered me. I wanted to understand the actual pipeline—not just the high-level theory, but the mechanical journey of the data.&lt;/p&gt;

&lt;p&gt;So, I visualized the life of a single prompt: "Write a poem about a robot."&lt;/p&gt;

&lt;p&gt;I traced it through Tokenization, Embeddings, Attention, and the KV Cache to understand how a matrix of numbers becomes a creative output.&lt;/p&gt;

&lt;p&gt;Here is the &lt;strong&gt;full visual breakdown&lt;/strong&gt; :&lt;/p&gt;

&lt;p&gt;

  &lt;iframe src="https://www.youtube.com/embed/x-XkExN6BkI"&gt;
  &lt;/iframe&gt;


&lt;/p&gt;

&lt;p&gt;🛠 T*&lt;em&gt;he Architecture Explained (Summary)&lt;/em&gt;*&lt;br&gt;
If you can't watch the video right now, here are the core mental models I used to make sense of the math.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Tokenization: The "Lego Brick" Phase&lt;/strong&gt;
The engine doesn't read English; it reads integers. Before anything happens, the tokenizer smashes our prompt into chunks.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;Input&lt;/strong&gt;: "Write a poem..."&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Output&lt;/strong&gt;: [1203, 45, 9001, ...]&lt;/p&gt;

&lt;p&gt;Think of these as Lego bricks. A simple word is one brick; a complex word might be three.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Embeddings: The "Hyper-Grocery Store"&lt;/strong&gt;
This was the biggest "Aha!" moment for me. How does the model know that "King" is related to "Queen"?&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;It's not a dictionary; it's a Grocery Store. In a grocery store, items aren't sorted alphabetically (Apples aren't next to Antifreeze). They are sorted by concept.&lt;/p&gt;

&lt;p&gt;Apples are near Bananas (Fruit aisle).&lt;/p&gt;

&lt;p&gt;Shampoo is near Soap (Hygiene aisle).&lt;/p&gt;

&lt;p&gt;The model converts our tokens into coordinates in a massive, multi-dimensional space. "Robot" isn't just a word; it's a vector located near "Metal," "Future," and "Technology."&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;The Attention Mechanism: The "Cocktail Party"&lt;/strong&gt;
This is the heavy lifting. Once the tokens are in the store, how do they relate to each other?&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;I visualized this as a Cocktail Party. Imagine you are at a loud party. You ignore 99% of the noise, but if someone shouts your name or a topic you love, you snap to attention.&lt;/p&gt;

&lt;p&gt;The model does exactly this. When processing the word "Bank," it looks back at the entire context window.&lt;/p&gt;

&lt;p&gt;If it sees the token "River," it pays attention to the "Nature" meaning of Bank.&lt;/p&gt;

&lt;p&gt;If it sees "Money," it pays attention to the "Finance" meaning.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;The Context Window: The "Carpenter's Workbench"&lt;/strong&gt;&lt;br&gt;
We often hear about context limits (8k, 32k, 128k). Think of the context window not as a brain, but as a physical workbench. You can only fit so many tools (tokens) on the bench at once. If you add too many, the oldest ones fall off the edge. This is why the model "hallucinates" or forgets things from the start of a long conversation—they literally fell off the table.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;RLHF: The Wolf vs. The Dog&lt;/strong&gt;&lt;br&gt;
Finally, I dug into why the model is polite. A raw base model (like GPT-3 before training) is a Wild Wolf. It just wants to hunt patterns. If you ask it a question, it might just ask you another question back, because that's what the training data looks like.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;RLHF (Reinforcement Learning from Human Feedback) is the process of domesticating that wolf into a helpful Labradoodle. We don't make the wolf smarter; we just train it to behave in a way that humans find useful.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Final Thoughts&lt;/strong&gt;&lt;br&gt;
Tracing the data path removed a lot of the "magic" for me, but it made me appreciate the engineering even more. It’s not a mind; it’s a probabilistic engine that is terrifyingly good at predicting the next Lego brick.&lt;/p&gt;

&lt;p&gt;If you want to see the animations for the KV Cache and Temperature, check out the video above!&lt;/p&gt;

&lt;p&gt;Let me know if these analogies click for you, or if you have a better way to visualize the Attention Mechanism! 👇&lt;/p&gt;

</description>
      <category>ai</category>
      <category>architecture</category>
      <category>llm</category>
      <category>chatgpt</category>
    </item>
    <item>
      <title>Code Review Visual Workflows using NotebookLM (The Video Method)</title>
      <dc:creator>PracticalAIGuy</dc:creator>
      <pubDate>Wed, 14 Jan 2026 22:58:48 +0000</pubDate>
      <link>https://dev.to/practicalaiguy_ba30448492/i-forced-google-notebooklm-to-roast-my-code-it-was-brutal-4625</link>
      <guid>https://dev.to/practicalaiguy_ba30448492/i-forced-google-notebooklm-to-roast-my-code-it-was-brutal-4625</guid>
      <description>&lt;p&gt;&lt;strong&gt;The "Yes-Man" Problem&lt;/strong&gt;&lt;br&gt;
We all know the struggle. You spend weekend building a complex AI agent. You stare at the code for so long that you lose all objectivity.&lt;/p&gt;

&lt;p&gt;If you ask ChatGPT, "Is this good?", it usually replies with excessive politeness: "Great job! This architecture is robust and scalable!"&lt;/p&gt;

&lt;p&gt;I didn't want a compliment. I wanted a Code Review. Actually, I wanted a Roast.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Experiment: Automated Red Teaming&lt;/strong&gt;&lt;br&gt;
I recently built an Autonomous Research Agent using n8n, OpenAI and Tavily (I wrote a full technical tutorial on how to build it here- &lt;a href="https://dev.to/practicalaiguy_ba30448492/i-built-an-ai-research-agent-to-cure-my-doomscrolling-addiction-42mi"&gt;https://dev.to/practicalaiguy_ba30448492/i-built-an-ai-research-agent-to-cure-my-doomscrolling-addiction-42mi&lt;/a&gt;).&lt;/p&gt;

&lt;p&gt;But before deploying it, I wanted to stress-test the logic. Since I didn't have a senior engineer handy to tear my code apart, I decided to build a synthetic one using &lt;strong&gt;Google NotebookLM&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Setup&lt;/strong&gt;&lt;br&gt;
NotebookLM's "&lt;strong&gt;Audio Overview&lt;/strong&gt;" feature is usually used for polite podcast summaries. But I discovered that if you use the &lt;strong&gt;Custom Instructions&lt;/strong&gt; feature, you can "jailbreak" the hosts into becoming hostile personas.&lt;/p&gt;

&lt;p&gt;I uploaded a video walkthrough of my n8n workflow and gave them this specific instruction:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The "Roast" Prompt&lt;/strong&gt;&lt;br&gt;
If you want to try this on your own projects, here is the exact prompt I used:&lt;/p&gt;

&lt;p&gt;Character: Act as two jaded, cynical tech reviewers.&lt;/p&gt;

&lt;p&gt;Tone:&lt;/p&gt;

&lt;p&gt;Be highly critical.&lt;/p&gt;

&lt;p&gt;Mock the complexity of the "no-code" graph.&lt;/p&gt;

&lt;p&gt;Use "hacker" slang (e.g., "spaghetti code", "wrapper").&lt;/p&gt;

&lt;p&gt;Narrative Arc:&lt;/p&gt;

&lt;p&gt;Start: Roast the project as just another "OpenAI wrapper."&lt;/p&gt;

&lt;p&gt;Middle: Notice the specific engineering details (specifically The Researcher - Tavily API to summarize news from multiple sources).&lt;/p&gt;

&lt;p&gt;End: Grudgingly admit that the architecture is actually valid and solid.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Result 💀&lt;/strong&gt;&lt;br&gt;
I expected a funny 30-second clip. What I got was a surprisingly accurate audit of my architecture.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The AI hosts:&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Mocked my "Visual Spaghetti": They correctly identified the chaos of my n8n canvas. They mocked, its one among the hundreds out there, but they regretted that comment later.&lt;/p&gt;

&lt;p&gt;Validated the Logic: By "roasting" it, they proved they actually understood why I built it that way.&lt;/p&gt;

&lt;p&gt;Found Value : Towards the end of the video, they found out additional value and use cases (which I never thought in the first place), and admit the tool is a highly customizable prototype.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Watch the Roast&lt;/strong&gt;&lt;br&gt;
You can hear the full audio (and see the agent breakdown) here:&lt;/p&gt;

&lt;p&gt;

  &lt;iframe src="https://www.youtube.com/embed/oof9JB3OFO4"&gt;
  &lt;/iframe&gt;


&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Why This is Actually Useful&lt;/strong&gt;&lt;br&gt;
Aside from the emotional damage, this is a legitimate productivity hack.&lt;/p&gt;

&lt;p&gt;When you are working solo, you live in an echo chamber. Forcing an LLM to adopt a hostile persona ("The Angry Senior Dev", "The Confused User", "The Security Auditor") breaks the "Yes-Man" loop.&lt;/p&gt;

&lt;p&gt;It forces the model to look for flaws instead of strengths.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Try it out&lt;/strong&gt;: Upload your own documentation or code video to NotebookLM, paste the prompt above, and see if your ego survives.&lt;/p&gt;

&lt;p&gt;Read the Build Guide: If you want to see the actual code and nodes that the AI was roasting, check out my Technical Build Tutorial here - &lt;a href="https://dev.to/practicalaiguy_ba30448492/i-built-an-ai-research-agent-to-cure-my-doomscrolling-addiction-42mi"&gt;https://dev.to/practicalaiguy_ba30448492/i-built-an-ai-research-agent-to-cure-my-doomscrolling-addiction-42mi&lt;/a&gt;.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>gemini</category>
      <category>roast</category>
      <category>productivity</category>
    </item>
    <item>
      <title>I Built an AI Research Agent to Cure My "Doomscrolling" Addiction</title>
      <dc:creator>PracticalAIGuy</dc:creator>
      <pubDate>Wed, 14 Jan 2026 00:10:18 +0000</pubDate>
      <link>https://dev.to/practicalaiguy_ba30448492/i-built-an-ai-research-agent-to-cure-my-doomscrolling-addiction-42mi</link>
      <guid>https://dev.to/practicalaiguy_ba30448492/i-built-an-ai-research-agent-to-cure-my-doomscrolling-addiction-42mi</guid>
      <description>&lt;p&gt;&lt;strong&gt;The Problem: AI News is Noise&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Every morning, I faced the same problem. There are 50 new AI tools released daily, 10 new models on HuggingFace, and endless hype on X/Twitter.&lt;/p&gt;

&lt;p&gt;I was wasting hours "doomscrolling" just to find the 2 or 3 updates that actually mattered to my work.&lt;/p&gt;

&lt;p&gt;I didn't need more news. I needed a &lt;strong&gt;Chief of Staff&lt;/strong&gt; to read everything for me, filter out the garbage, and only show me the signal.&lt;/p&gt;

&lt;p&gt;So, I built one.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Solution: An Autonomous "News Editor"&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;In this tutorial, I’ll show you how I built a &lt;strong&gt;Personal AI News Agent using n8n, OpenAI, and Tavily&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;It works while I sleep:&lt;/p&gt;

&lt;p&gt;**Reads **the raw RSS feeds from major tech sites.&lt;/p&gt;

&lt;p&gt;**Judges **every headline (acting as a strict "Senior Editor").&lt;/p&gt;

&lt;p&gt;**Researches **the winners using Tavily (to verify facts).&lt;/p&gt;

&lt;p&gt;**Delivers **a curated morning briefing to my email.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Stack&lt;/strong&gt;&lt;br&gt;
&lt;strong&gt;Orchestrator&lt;/strong&gt;: n8n (Local or Cloud).&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Brain&lt;/strong&gt; (Filter): OpenAI gpt-4o-mini (Cheap and fast).&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Researcher&lt;/strong&gt;: Tavily AI (Essential for fetching live context).&lt;/p&gt;
&lt;h2&gt;
  
  
  &lt;strong&gt;Source&lt;/strong&gt;: RSS Feeds (e.g., TechCrunch, Verge).
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Step 1: The "Firehose" (RSS Ingestion)&lt;/strong&gt;&lt;br&gt;
The workflow starts with a &lt;strong&gt;Schedule Trigger&lt;/strong&gt; set for 8:00 AM. It pulls the latest articles using the &lt;strong&gt;RSS Read Node&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;At this stage, we have everything—rumors, minor updates, and noise.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 2: The "Senior Editor" (OpenAI Filtering)&lt;/strong&gt;&lt;br&gt;
This is the most critical part. I didn't just ask AI to "summarize." I used a Loop Node to process each headline individually and gave OpenAI a specific persona:&lt;/p&gt;

&lt;p&gt;_System Prompt: "Analyze this news item:&lt;br&gt;
Title: {{ $json.title }}&lt;br&gt;
Summary: {{ $json.contentSnippet || $json.content }}&lt;/p&gt;

&lt;p&gt;YOUR ROLE:&lt;br&gt;
You are a Senior Tech Editor curating a daily briefing. Your goal is to identify useful, relevant news for AI Engineers.&lt;/p&gt;

&lt;p&gt;SCORING GUIDELINES (0-10):&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;0-3: Irrelevant, gossip, or low-quality clickbait.&lt;/li&gt;
&lt;li&gt;4-5: Average news. Minor updates or generic articles.&lt;/li&gt;
&lt;li&gt;6-7 (PASSING): Solid, useful news. Good tutorials, interesting tool releases, or standard industry updates.&lt;/li&gt;
&lt;li&gt;8-10 (EXCELLENT): Major breakthroughs, aquistitions, critical security alerts, or high-impact releases (e.g., GPT-5, new SOTA model).&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;INSTRUCTIONS:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Rate strictly but fairly.&lt;/li&gt;
&lt;li&gt;If it is useful to a professional, give it at least a 6.&lt;/li&gt;
&lt;li&gt;Return ONLY a JSON object.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;OUTPUT FORMAT:&lt;br&gt;
{&lt;br&gt;
  "score": ,&lt;br&gt;
   "title": &lt;/p&gt;
,&lt;br&gt;
  "reason": ""&lt;br&gt;
}_

&lt;p&gt;&lt;strong&gt;Step 3: The Gatekeeper (If Node)&lt;/strong&gt;&lt;br&gt;
I added an If Node that acts as a gate.&lt;/p&gt;

&lt;p&gt;_Score &amp;lt; 7: Discard immediately.&lt;/p&gt;

&lt;p&gt;Score &amp;gt;= 7: Proceed to research._&lt;/p&gt;

&lt;p&gt;This simple logic reduced my reading list from ~50 articles to just the top 5.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 4: The Deep Dive (Tavily AI)&lt;/strong&gt;&lt;br&gt;
For the winning articles, I didn't want just the RSS blurb. I used Tavily AI to go out and "read" the full context of the story.&lt;/p&gt;

&lt;p&gt;I set Tavily's &lt;strong&gt;include_answer&lt;/strong&gt; parameter to "&lt;strong&gt;Advanced&lt;/strong&gt;." This generates a high-quality, synthesized summary of the topic based on multiple sources, not just the original article.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 5: The Briefing (Email)&lt;/strong&gt;&lt;br&gt;
Finally, an Aggregate Node collects all the "Winners" and formats them into a clean HTML email, sent via Gmail.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi3mmuwstv76aswytrzop.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi3mmuwstv76aswytrzop.png" alt=" "&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Watch the Build (Step-by-Step)&lt;br&gt;
I recorded the entire process, including the exact Prompt and JSON logic I used. You can follow along here:&lt;/p&gt;

&lt;p&gt;

  &lt;iframe src="https://www.youtube.com/embed/mOnbK6DuFhc"&gt;
  &lt;/iframe&gt;


&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Why This Matters&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;By building this agent, I saved myself ~5 hours a week of mindless scrolling. The agent does the boring work of filtering; I just read the high-signal results.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Next Steps&lt;/strong&gt;: In my next post, I’ll share how I used Google NotebookLM to "stress test" this agent.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Let me know in the comments&lt;/strong&gt;: How are you handling the information overload right now?&lt;/p&gt;

</description>
      <category>ai</category>
      <category>automation</category>
      <category>tavily</category>
      <category>openai</category>
    </item>
  </channel>
</rss>
