<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Kalpana R</title>
    <description>The latest articles on DEV Community by Kalpana R (@kalpanait161).</description>
    <link>https://dev.to/kalpanait161</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F704350%2Ff401ad00-120c-4ac9-906a-e912a59235c5.png</url>
      <title>DEV Community: Kalpana R</title>
      <link>https://dev.to/kalpanait161</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/kalpanait161"/>
    <language>en</language>
    <item>
      <title>🎓 Session 1: Hello World of RAG + Introduction &amp; Need of RAG</title>
      <dc:creator>Kalpana R</dc:creator>
      <pubDate>Wed, 29 Apr 2026 07:24:00 +0000</pubDate>
      <link>https://dev.to/kalpanait161/session-1-hello-world-of-rag-introduction-need-of-rag-54al</link>
      <guid>https://dev.to/kalpanait161/session-1-hello-world-of-rag-introduction-need-of-rag-54al</guid>
      <description>&lt;p&gt;&lt;strong&gt;What I Learned&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Today’s session introduced me to &lt;strong&gt;Retrieval-Augmented Generation (RAG)&lt;/strong&gt; and why it’s becoming essential in AI. The focus was on understanding the &lt;strong&gt;limitations of plain language models (LLMs)&lt;/strong&gt; and how RAG helps overcome them.&lt;br&gt;
..&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What is RAG?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Retrieval-Augmented Generation (RAG)&lt;/strong&gt; is a technique that combines:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Retrieval&lt;/strong&gt; → fetching relevant, external information (from documents, databases, or the web).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Generation&lt;/strong&gt; → using a language model (LLM) to produce coherent, context-aware responses.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The term “&lt;strong&gt;Augmented&lt;/strong&gt;” means the LLM’s generation process is enhanced by the retrieved information — instead of relying only on its internal training data, it’s augmented with fresh, factual context from external sources.&lt;/p&gt;

&lt;p&gt;Together, RAG helps produce outputs that are more &lt;strong&gt;factual, relevant, and up-to-date&lt;/strong&gt; by grounding responses in retrieved information.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fl72d8nsbzgx0368rq2df.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fl72d8nsbzgx0368rq2df.png" alt=" " width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;&lt;strong&gt;Key Concepts:&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Language Models (LLMs)&lt;/strong&gt; generate text by predicting the next word based on learned patterns and context.&lt;br&gt;
They use probabilities and context to produce coherent and meaningful responses.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Limitations of LLMs:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Hallucinations (making up answers when unsure).&lt;/li&gt;
&lt;li&gt;Outdated knowledge (training data has a cutoff).&lt;/li&gt;
&lt;li&gt;No access to private or domain-specific documents.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;RAG:&lt;/strong&gt; Combines retrieval (fetching relevant info) with generation (LLM output).&lt;br&gt;
     - RAG helps improve factuality, relevance, and freshness by grounding responses in retrieved information&lt;/p&gt;




&lt;p&gt;&lt;strong&gt;How Do LLMs Learn? (Weights &amp;amp; Parameters)&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The language models are like equations with parameters. During training, the model adjusts its &lt;strong&gt;weights&lt;/strong&gt; — internal values that decide how strongly one word or feature influences another.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Example: Just as changing 𝑚 or 𝑐 in 𝑦=𝑚𝑥+𝑐 changes the line, adjusting weights changes the model’s predictions.&lt;/li&gt;
&lt;li&gt;These weights are what allow the model to learn patterns from massive datasets.&lt;/li&gt;
&lt;/ul&gt;




&lt;p&gt;&lt;strong&gt;SLM vs. LLM (General Note)&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;While this session focused mainly on Large Language Models (LLMs) and RAG, it’s useful to know the distinction:&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;LLMs (Large Language Models)&lt;/strong&gt; → Very big models trained on massive datasets. They’re powerful, but resource-heavy.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;SLMs (Small Language Models)&lt;/strong&gt; → More compact models designed for efficiency. They can run faster, use less memory, and are easier to deploy on devices with limited resources.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;In practice:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;LLMs are great for complex reasoning and broad knowledge.&lt;/li&gt;
&lt;li&gt;SLMs are often used for lightweight tasks, edge devices, or situations where speed and efficiency matter.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This is useful context to keep in mind as I continue learning about RAG and AI systems.&lt;/p&gt;




&lt;p&gt;&lt;strong&gt;Why Do We Need RAG?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Plain language models are powerful but limited:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;They hallucinate (make up answers when unsure).&lt;/li&gt;
&lt;li&gt;They rely on static training data (no updates after cutoff).&lt;/li&gt;
&lt;li&gt;They can’t access private or domain-specific documents.
RAG helps reduce these issues by grounding answers in retrieved context.&lt;/li&gt;
&lt;/ul&gt;




&lt;p&gt;&lt;strong&gt;Key Examples from the Session&lt;/strong&gt; &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;Dogs, Cats, and Lion&lt;/strong&gt; →&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Without RAG:&lt;/strong&gt; If a model has not seen enough relevant information about lions in its training data, it may generate incorrect or fabricated answers (hallucinations).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;With RAG:&lt;/strong&gt; Retrieval brings in factual information about lions from external sources, helping the model generate a more accurate and grounded response.&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;

&lt;p&gt;&lt;strong&gt;COVID vs. Current Events&lt;/strong&gt; →&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Without RAG:&lt;/strong&gt; The model may know about COVID (from training data) but struggle with recent events due to outdated knowledge.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;With RAG:&lt;/strong&gt; Retrieval pulls in recent articles or documents, allowing the model to respond with up-to-date context.&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;

&lt;p&gt;&lt;strong&gt;River Bank&lt;/strong&gt; → Context confusion: “river bank” vs. “financial bank.”&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Without RAG:&lt;/strong&gt; The model may confuse “river bank” (geography) with “bank” (finance) depending on context.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;With RAG:&lt;/strong&gt; Retrieval provides relevant domain context, helping the model choose the correct meaning.&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;

&lt;p&gt;&lt;strong&gt;Company Docs&lt;/strong&gt; → LLM alone can’t answer from private files, but RAG can.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Without RAG:&lt;/strong&gt; The model cannot access private or internal company documents.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;With RAG:&lt;/strong&gt; Retrieval fetches relevant internal documents, enabling accurate answers based on company data.&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;

&lt;p&gt;&lt;strong&gt;Hello Predictions&lt;/strong&gt;→ &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Without RAG&lt;/strong&gt;: With “Hello,” low temperature may produce “World,” while high temperature may produce “How are you?” or other creative outputs — but answers may drift.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;With RAG&lt;/strong&gt;: Even at high temperature, retrieval keeps outputs grounded in factual context.&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;/ul&gt;




&lt;p&gt;&lt;strong&gt;Temperature Settings&lt;/strong&gt; &lt;strong&gt;[Temperature in LLMs]&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Low Temperature (~0)&lt;/strong&gt; → More deterministic and consistent responses.&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;High Temperature (~1 or above)&lt;/strong&gt; → More creative and varied responses.&lt;br&gt;
-&lt;strong&gt;Takeaway:&lt;/strong&gt; Use low temperature for consistency and high temperature for creativity. Note that temperature controls randomness, not correctness.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;My Note: Retrieval can help guide responses with relevant context, even when temperature increases variability.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;




&lt;p&gt;&lt;strong&gt;Real-World Applications of RAG&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Customer Support&lt;/strong&gt; → Answers from FAQs and manuals.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Healthcare&lt;/strong&gt; → Grounded responses from medical databases.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Education&lt;/strong&gt; → Fact-checked explanations for learners.&lt;/li&gt;
&lt;li&gt;*&lt;em&gt;Enterprise Search *&lt;/em&gt;→ Unlocking insights from private organizational data.&lt;/li&gt;
&lt;/ul&gt;




&lt;p&gt;&lt;strong&gt;Key Takeaways (Quick Reference)&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;RAG = Retrieval + Generation.&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt; Helps reduce hallucinations, outdated knowledge issues, and lack of private context.&lt;/li&gt;
&lt;li&gt;Temperature controls creativity vs. accuracy.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Real-world uses:&lt;/strong&gt; support, healthcare, education, enterprise search.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Core idea:&lt;/strong&gt; ground AI in facts before generating answers.&lt;/li&gt;
&lt;/ul&gt;




&lt;p&gt;&lt;strong&gt;My Conclusion&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Today’s session gave me a strong foundation in understanding the &lt;strong&gt;limitations of AI&lt;/strong&gt; and how &lt;strong&gt;RAG helps overcome them.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Instead of relying only on memory, RAG allows AI to &lt;strong&gt;look up relevant information before answering&lt;/strong&gt;—just like how we perform better when we can refer to notes.&lt;/p&gt;

&lt;p&gt;This is just the beginning of my learning journey with RAG — I’ll continue documenting as I go.&lt;/p&gt;




&lt;p&gt;📚 This post is part of my &lt;em&gt;Learning Notes – RAG Series&lt;/em&gt;.&lt;br&gt;&lt;br&gt;
Next up: Session 2, where I’ll continue exploring and documenting my journey.&lt;/p&gt;

</description>
      <category>rag</category>
      <category>beginners</category>
      <category>llm</category>
      <category>learningjourney</category>
    </item>
  </channel>
</rss>
