DEV Community

# llm

Posts

👋 Sign in for the ability to sort posts by relevant, latest, or top.
RAG in the Wild: What I Learned After Two Weeks of Chunking Experiments

RAG in the Wild: What I Learned After Two Weeks of Chunking Experiments

Comments 2
7 min read
How to Reduce OpenAI Bill Without Hurting Quality: A Practical Audit Framework

How to Reduce OpenAI Bill Without Hurting Quality: A Practical Audit Framework

6
Comments 3
6 min read
Monitor LLM Inference in Production (2026): Prometheus & Grafana for vLLM, TGI, llama.cpp

Monitor LLM Inference in Production (2026): Prometheus & Grafana for vLLM, TGI, llama.cpp

1
Comments
9 min read
I Read a Paper That Genuinely Made Me Stop and Think — AI is Now Jailbreaking Other AI

I Read a Paper That Genuinely Made Me Stop and Think — AI is Now Jailbreaking Other AI

Comments
3 min read
Your AI Gateway Just Became an Attack Vector: Anatomy of the LiteLLM Supply Chain Compromise

Your AI Gateway Just Became an Attack Vector: Anatomy of the LiteLLM Supply Chain Compromise

1
Comments 1
7 min read
Why Your RAG System Returns Garbage (And How to Actually Fix It)

Why Your RAG System Returns Garbage (And How to Actually Fix It)

Comments
5 min read
Six Characters Fixed My AI's Personality: A Fine-Tuning Story

Six Characters Fixed My AI's Personality: A Fine-Tuning Story

Comments
4 min read
Why Your AI Agents Are Burning Cash and How to Fix It

Why Your AI Agents Are Burning Cash and How to Fix It

Comments
5 min read
Llama vs Mistral vs Phi: Complete Open-Source LLM Comparison for Enterprise (2026)

Llama vs Mistral vs Phi: Complete Open-Source LLM Comparison for Enterprise (2026)

Comments
16 min read
Introducing llm-lean-log: Token-Efficient Chat Logging for AI Agents

Introducing llm-lean-log: Token-Efficient Chat Logging for AI Agents

Comments
4 min read
LLMs - How Did They Get So Good?

LLMs - How Did They Get So Good?

4
Comments 2
13 min read
Revolutionary LLM‑Generated Helm Charts: Build, Test, Deploy in Minutes

Revolutionary LLM‑Generated Helm Charts: Build, Test, Deploy in Minutes

1
Comments
5 min read
Fine-tuning vs RAG: Cuándo Usar Cada Enfoque para LLMs en Producción

Fine-tuning vs RAG: Cuándo Usar Cada Enfoque para LLMs en Producción

Comments
8 min read
vLLM vs SGLang vs LMDeploy: Fastest LLM Inference Engine in 2026?

vLLM vs SGLang vs LMDeploy: Fastest LLM Inference Engine in 2026?

1
Comments
9 min read
Fine-tuning vs RAG: Cuándo Usar Cada Enfoque para LLMs en Producción

Fine-tuning vs RAG: Cuándo Usar Cada Enfoque para LLMs en Producción

Comments
8 min read
👋 Sign in for the ability to sort posts by relevant, latest, or top.