This is a submission for the AI Agents Challenge powered by n8n and Bright Data
What I Built
Online communities like Reddit are a treasure trove of user concerns, struggles, and honest feedback. Every day, people share their unfiltered experiences—whether it’s about skincare routines, tech products, mental health, or niche hobbies. For businesses, researchers, or even curious individuals, this raw conversation is a window into what people actually think, feel, and need.
But the problem? These discussions are often chaotic, unstructured, and scattered across countless comments—making it overwhelming to extract actionable insights.
That’s where my project comes in.
I built an AI-powered workflow that:
Analyzes Reddit posts and comments in real time
Extracts the underlying concerns, themes, and trends
Automatically generates clear, structured reports inside Notion
In short, it transforms messy threads into digestible knowledge bases—perfect for:
Marketers → tracking sentiment, product pain points, emerging trends
Researchers → organizing authentic, large-scale community insights
Everyday users → turning Reddit rabbit holes into useful, structured knowledge
This workflow bridges the gap between community chatter and real understanding, making online discussions finally usable as living, actionable data.
Demo
n8n Workflow
Technical Implementation
This project leverages n8n’s modular workflow engine, Bright Data’s Verified Node, and Google Gemini to build a real-time Reddit insights agent that scrapes posts + comments, processes them, and delivers structured marketing insights into Notion.
Here’s a breakdown of the key nodes and how they fit into the workflow:
Schedule Trigger: Runs the workflow on a timed basis, ensuring continuous monitoring of subreddit activity.
Set Subreddit: Defines which subreddit to analyze (hardcoded or user-input).
RSS Read: Pulls the most recent posts from the chosen subreddit’s RSS feed.
Extract URLs → Aggregate URLs: Cleans and batches the post URLs to prepare them for Bright Data’s scraper.
Bright Data Nodes: (see next section for full detail) Used to initiate, monitor, and download data-rich snapshots (post details, comments, upvotes).
Loop Over Items & Control Flow: Ensures that batch extractions are retried until Bright Data reports readiness.
Data Preparation (Set & Aggregate): Extracts essential fields (title, description, comments, upvotes, etc.) and reduces them into a single object for LLM input.
AI Agent Node: Uses Google Gemini as the underlying model, with system prompts tailored to generate actionable marketing insights in structured JSON format.
Notion Integration:
Create a page → initializes a new Notion page for each insights report.
Md to Notion Blocks → converts AI output (Markdown) into Notion-compatible blocks.
Add blocks as Children → appends the AI-generated insights into the Notion workspace.
This combination allows the workflow to run end-to-end autonomously: from subreddit scraping → to AI-driven summarization → to real-time reporting in Notion.
Bright Data Verified Node
Bright Data’s Verified Node was the data backbone of this project. Here’s how it was used step-by-step:
Initiate Batch Extraction from URL
Operation: triggerCollectionByUrl
Input: A batched list of subreddit post URLs.
Output: A snapshot_id representing the batch extraction job.
Purpose: Kickstarts Bright Data’s web scraper to collect post metadata and comments.
Check the Status of Batch Extraction
Operation: monitorProgressSnapshot
Input: snapshot_id
Output: Status values (running, ready, etc.).
Purpose: Polls Bright Data until the dataset is marked as ready.
Control Flow: Loop & Condition
Loop Over Items + Wait 30 Seconds: Creates a polling mechanism.
If Check → Batch Ready: Ensures data is only downloaded when Bright Data confirms completion.
Download the Snapshot Content
Operation: downloadSnapshot
Input: snapshot_id
Output: Full structured dataset (title, description, comments, replies, votes, etc.).
Purpose: Provides the rich Reddit dataset that powers the AI analysis.
Post-processing (Extract Essential Data)
Fields captured: title, description, comments, upvotes, num_comments, url.
Structured into a clean JSON object for downstream AI consumption.
The Bright Data integration ensures that the AI isn’t relying on static or partial information—it has access to real-time, community-level conversations with full context, making the insights both fresh and actionable.
Journey
The biggest challenge I faced was converting the AI Agent Node’s output into properly formatted Notion documents.
I had to dive deep into the Notion API — learning about blocks, page IDs, and content structure — before I could get everything working smoothly.
This process gave me valuable hands-on experience connecting AI outputs with real productivity tools.
For the future, I plan to build a more detailed workflow and documentation so others can easily replicate and extend it.
Future Improvements
- Multi-Subreddit Monitoring – Track discussions across multiple subreddits to uncover community-wide trends.
- Notion Database Integration – Push insights directly into databases for better querying, filtering, and reporting (beyond just static pages).
- Advanced Filtering – Add layers for sentiment analysis, product mentions, and brand signal detection to deliver sharper insights.
- Cross-Platform Export – Support sending reports not only to Notion but also to Google Docs, Slack, Email, or other productivity tools.
- Setup Guide for the Community – Publish a detailed tutorial and documentation so anyone can replicate or customize the workflow.
-
Towards a SaaS Product – Transform this into a user-facing platform where:
- Users can securely input their Bright Data key and Notion credentials.
- Select multiple subreddits (or future sources like Twitter/X, Discord, or forums).
- Choose preferred output channels (Notion, Google Docs, Slack, Email, etc.).
- View insights directly on the web app or integrate with their existing stack.
Vision Statement:
My long-term vision is to build a SaaS platform that democratizes community insights — so anyone, from solo creators to product teams, can tap into real-time online conversations and instantly turn them into structured, actionable knowledge without needing technical setup.
With n8n + Bright Data + AI, I’ve shown how to turn chaotic online conversations into structured, actionable intelligence — directly inside Notion 🔥.
Top comments (0)