<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Matt Engman</title>
    <description>The latest articles on DEV Community by Matt Engman (@matt_engman_a63bbe68e2498).</description>
    <link>https://dev.to/matt_engman_a63bbe68e2498</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F2971521%2Fba490e36-85f6-4d68-9c49-dffee80f4eb1.jpg</url>
      <title>DEV Community: Matt Engman</title>
      <link>https://dev.to/matt_engman_a63bbe68e2498</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/matt_engman_a63bbe68e2498"/>
    <language>en</language>
    <item>
      <title>DocuMind - Production-Ready Semantic Document Search with Redis 8 Vector Sets</title>
      <dc:creator>Matt Engman</dc:creator>
      <pubDate>Fri, 01 Aug 2025 17:01:41 +0000</pubDate>
      <link>https://dev.to/matt_engman_a63bbe68e2498/documind-production-ready-semantic-document-search-with-redis-8-vector-sets-3dbi</link>
      <guid>https://dev.to/matt_engman_a63bbe68e2498/documind-production-ready-semantic-document-search-with-redis-8-vector-sets-3dbi</guid>
      <description>&lt;p&gt;&lt;em&gt;This is a submission for the &lt;a href="https://dev.to/challenges/redis-2025-07-23"&gt;Redis AI Challenge&lt;/a&gt;: Real-Time AI Innovators&lt;/em&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  What I Built
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;DocuMind&lt;/strong&gt; is a production-ready semantic document search system that transforms static document storage into an intelligent, searchable knowledge base. Built with Redis 8 Vector Sets at its core, DocuMind enables natural language queries across entire document collections with sub-second response times.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Key Features:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;🔍 &lt;strong&gt;Real-time semantic search&lt;/strong&gt; using OpenAI embeddings and Redis Vector Sets&lt;/li&gt;
&lt;li&gt;📊 &lt;strong&gt;Live analytics dashboard&lt;/strong&gt; with search metrics and system health monitoring&lt;/li&gt;
&lt;li&gt;📄 &lt;strong&gt;Intelligent document processing&lt;/strong&gt; with automatic chunking and vector generation&lt;/li&gt;
&lt;li&gt;⚡ &lt;strong&gt;Advanced caching&lt;/strong&gt; with 75% memory efficiency through quantized embeddings&lt;/li&gt;
&lt;li&gt;🎯 &lt;strong&gt;Production deployment&lt;/strong&gt; on Google Cloud Run + Vercel with enterprise security&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Tech Stack:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Backend&lt;/strong&gt;: FastAPI, Redis 8, OpenAI API, Python&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Frontend&lt;/strong&gt;: React, TypeScript, Tailwind CSS, Framer Motion&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Infrastructure&lt;/strong&gt;: Google Cloud Run, Vercel, Redis Cloud&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AI/ML&lt;/strong&gt;: OpenAI embeddings, sentence-transformers fallback&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Demo
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://documind-ruby.vercel.app/" rel="noopener noreferrer"&gt;🚀 &lt;strong&gt;Live Demo&lt;/strong&gt;:&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Try it yourself:&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Upload a PDF, DOCX, or TXT document&lt;/li&gt;
&lt;li&gt;Watch real-time processing with vector generation&lt;/li&gt;
&lt;li&gt;Search using natural language (e.g., "artificial intelligence", "business strategy")&lt;/li&gt;
&lt;li&gt;View live analytics showing search performance and system metrics&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;a href="https://github.com/MatthewEngman/documind" rel="noopener noreferrer"&gt;&lt;strong&gt;GitHub Repository&lt;/strong&gt;:&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  How I Used Redis 8
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;DocuMind leverages Redis 8 as the foundation for its entire real-time AI infrastructure:&lt;/strong&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  🎯 &lt;strong&gt;Redis Vector Sets - The Core Innovation&lt;/strong&gt;
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Native vector storage&lt;/strong&gt; using Redis 8's cutting-edge Vector Sets feature&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Optimized fallback search&lt;/strong&gt; when Redis Stack KNN queries aren't available&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Base64 vector encoding&lt;/strong&gt; for reliable storage and retrieval&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Quantized embeddings&lt;/strong&gt; achieving 75% memory reduction vs traditional databases&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  📊 &lt;strong&gt;Multi-Model Data Architecture&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;DocuMind uses Redis 8's versatility to store multiple data types seamlessly:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;JSON documents&lt;/strong&gt; for metadata and document information&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Vector Sets&lt;/strong&gt; for semantic embeddings and similarity search&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Hash maps&lt;/strong&gt; for analytics and system metrics&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Sets&lt;/strong&gt; for document indexing and relationship management&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  ⚡ &lt;strong&gt;Real-Time Performance Features&lt;/strong&gt;
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Semantic caching&lt;/strong&gt; with intelligent cache invalidation&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Sub-second search&lt;/strong&gt; across thousands of document chunks&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Live analytics&lt;/strong&gt; tracking search patterns and system health&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Background processing&lt;/strong&gt; with Redis-based job queues&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  🔧 &lt;strong&gt;Production-Grade Implementation&lt;/strong&gt;
&lt;/h3&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;
python
# Redis Vector Sets integration with fallback search
async def search_vectors(self, query: str, limit: int = 10, 
                        similarity_threshold: float = 0.1):
    # Generate OpenAI embedding
    query_embedding = await embedding_service.generate_embedding(query)
    query_vector = np.array(query_embedding["vector"], dtype=np.float32)

    # Use optimized fallback vector search for Redis Stack compatibility
    results = await self._execute_fallback_search(query_vector, limit)

    # Process and rank results with cosine similarity
    return self._process_search_results(results, query_vector, similarity_threshold)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

</description>
      <category>redischallenge</category>
      <category>devchallenge</category>
      <category>database</category>
      <category>ai</category>
    </item>
  </channel>
</rss>
