DEV Community

# rag

Retrieval augmented generation, or RAG, is an architectural approach that can improve the efficacy of large language model (LLM) applications by leveraging custom data.

Posts

👋 Sign in for the ability to sort posts by relevant, latest, or top.
The Future of Hyper-Local AI

The Future of Hyper-Local AI

Comments 1
1 min read
Building Vroom AI: A Multi-Agent Architecture for Intelligent Driving Education

Building Vroom AI: A Multi-Agent Architecture for Intelligent Driving Education

3
Comments
7 min read
10 Best Practices to Manage Unstructured Data for Enterprises

10 Best Practices to Manage Unstructured Data for Enterprises

Comments
8 min read
Building a Local-First RAG Engine for AI Coding Assistants

Building a Local-First RAG Engine for AI Coding Assistants

Comments
4 min read
Self-Hosting Cognee: LLM Performance Tests

Self-Hosting Cognee: LLM Performance Tests

Comments
9 min read
Clone Your CTO: The Architecture of an 'AI Twin' (DSPy + Unsloth)

Clone Your CTO: The Architecture of an 'AI Twin' (DSPy + Unsloth)

Comments
3 min read
How I Improved RAG Accuracy from 73% to 100% - A Chunking Strategy Comparison

How I Improved RAG Accuracy from 73% to 100% - A Chunking Strategy Comparison

Comments
7 min read
Enterprise-Grade RAG Platform: Orchestrating Amazon Bedrock Agents via Red Hat OpenShift AI

Enterprise-Grade RAG Platform: Orchestrating Amazon Bedrock Agents via Red Hat OpenShift AI

Comments
22 min read
One Year of Model Context Protocol: From Experiment to Industry Standard

One Year of Model Context Protocol: From Experiment to Industry Standard

Comments
3 min read
TOON vs JSON en RAG (Java): el Grinch de los formatos cuando cada token cuenta 🎁

TOON vs JSON en RAG (Java): el Grinch de los formatos cuando cada token cuenta 🎁

Comments
7 min read
Modern Search Techniques for Vector Databases (w/LangChain)

Modern Search Techniques for Vector Databases (w/LangChain)

1
Comments
4 min read
The Research: MiniMax M2.1 (The "Linear" Revolution)

The Research: MiniMax M2.1 (The "Linear" Revolution)

Comments
3 min read
Deploying Scalable LLM Tools via Remote MCP on Kubernetes

Deploying Scalable LLM Tools via Remote MCP on Kubernetes

Comments
10 min read
Build a Knowledge-Based Q&A Bot using Bedrock + S3 + DynamoDB/OpenSearch via AWS CDK

Build a Knowledge-Based Q&A Bot using Bedrock + S3 + DynamoDB/OpenSearch via AWS CDK

Comments
25 min read
VectorDatabase Showdown 2025: Pinecone vs Qdrant vs Weaviate con Benchmarks Reales

VectorDatabase Showdown 2025: Pinecone vs Qdrant vs Weaviate con Benchmarks Reales

Comments
3 min read
The cheapest way to make agents reliable: define scope like a contract (not a vibe)

The cheapest way to make agents reliable: define scope like a contract (not a vibe)

Comments
4 min read
Graph RAG: Why Vector Search Alone Is Not Enough for Serious Backend Systems

Graph RAG: Why Vector Search Alone Is Not Enough for Serious Backend Systems

Comments
2 min read
Build a RAG Pipeline with n8n: Visual Workflows vs. Code-First

Build a RAG Pipeline with n8n: Visual Workflows vs. Code-First

Comments
6 min read
If your agent can delete user data, your prompt isn’t a prompt, it’s a contract

If your agent can delete user data, your prompt isn’t a prompt, it’s a contract

Comments 2
4 min read
RAG Without the Internet - Lessons From Building an Internal-Only AI Assistant on Markdown and Confluence

RAG Without the Internet - Lessons From Building an Internal-Only AI Assistant on Markdown and Confluence

Comments
6 min read
Building a RAG based agent using DronaHQ

Building a RAG based agent using DronaHQ

Comments
8 min read
AI Agents Feel Simple in Demos and Complicated in Production

AI Agents Feel Simple in Demos and Complicated in Production

Comments
2 min read
Building Memory for AI-Assisted Development

Building Memory for AI-Assisted Development

1
Comments
5 min read
The Tradeoffs Behind AI Agents

The Tradeoffs Behind AI Agents

Comments
1 min read
Optimizing Milvus Standalone for Production: Achieving 72% Memory Reduction While Maintaining Performance

Optimizing Milvus Standalone for Production: Achieving 72% Memory Reduction While Maintaining Performance

Comments
3 min read
loading...