Keeping up with AI news means scraping the same sources as everyone else — HackerNews, Reddit (r/MachineLearning, r/LocalLLaMA), ArXiv, GitHub Trending, Dev.to, Lobsters, and about 10 more. Everyone builds their own version. Seems silly.
So we built one and open sourced it.
📚 Full writeup: https://treesoop.com/blog/ai-news-mcp-17-sources-auto-scraping-2026
🔧 GitHub: https://github.com/treesoop/ai-news-mcp
What it does
17 sources scraped every 6 hours by a Mac mini in our office. Results exposed via Model Context Protocol so any MCP-compatible AI tool can query it.
Sources include:
- HackerNews, Reddit (4 AI subs), Dev.to, Lobsters
- ArXiv AI, ArXiv ML
- GitHub Trending
- OpenAI, Anthropic, Google AI, Meta AI blogs
- TechCrunch AI, VentureBeat AI, The Verge AI
Install
claude mcp add ai-news -- npx -y @treesoop/ai-news-mcp
Works in Claude Code, Cursor, Claude Desktop, ChatGPT — anywhere MCP is supported.
Example queries
- "Show me HackerNews top AI posts from today with 100+ points"
- "Summarize ArXiv AI papers about RAG from the last 24 hours"
- "What's trending on r/LocalLLaMA about Qwen3?"
Why bother
Our blog agent at treesoop.com uses this MCP to decide what to write about each day. Before it existed, we were scraping manually in 5 different scripts. Now it's one call.
If you run a dev newsletter, slack bot, or content agent, this probably saves you an afternoon.
Details
- MIT licensed, commercial use OK
- Data stored in our Supabase instance (free for public use)
- Self-host option in README if you want your own cadence
More TreeSoop OSS: hwp-mcp, whisper_transcription, claude2codex.
Top comments (0)