<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Eze Lanza</title>
    <description>The latest articles on DEV Community by Eze Lanza (@eze_lanza).</description>
    <link>https://dev.to/eze_lanza</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1390346%2F4ed56bb8-0578-42f6-97a5-536f940db0a9.jpg</url>
      <title>DEV Community: Eze Lanza</title>
      <link>https://dev.to/eze_lanza</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/eze_lanza"/>
    <language>en</language>
    <item>
      <title>Which IDE do you use? Cursor / Codeium / Zed / Fleet ?</title>
      <dc:creator>Eze Lanza</dc:creator>
      <pubDate>Thu, 13 Mar 2025 19:06:19 +0000</pubDate>
      <link>https://dev.to/eze_lanza/which-ide-do-you-use-cursor-codeium-zed-fleet--em4</link>
      <guid>https://dev.to/eze_lanza/which-ide-do-you-use-cursor-codeium-zed-fleet--em4</guid>
      <description></description>
      <category>discuss</category>
    </item>
    <item>
      <title>¿Quieres aprender sobre agentes en español? 🎥</title>
      <dc:creator>Eze Lanza</dc:creator>
      <pubDate>Thu, 13 Mar 2025 18:53:08 +0000</pubDate>
      <link>https://dev.to/eze_lanza/quieres-aprender-sobre-agentes-en-espanol-1aop</link>
      <guid>https://dev.to/eze_lanza/quieres-aprender-sobre-agentes-en-espanol-1aop</guid>
      <description>&lt;p&gt;He creado este video como parte de una serie de entrenamientos gratuitos para la comunidad latina. ¡Espero que te sea útil! 🚀&lt;/p&gt;

&lt;p&gt;🔗 Míralo aquí:&lt;/p&gt;

&lt;p&gt;&lt;iframe width="710" height="399" src="https://www.youtube.com/embed/yDzYesxqRdg"&gt;
&lt;/iframe&gt;
&lt;/p&gt;

</description>
      <category>llm</category>
      <category>rag</category>
      <category>ai</category>
      <category>tutorial</category>
    </item>
    <item>
      <title>¿Quieres aprender sobre agentes en español? 🎥 He creado este video como parte de una serie de entrenamientos gratuitos para la comunidad latina. ¡Espero que https://www.youtube.com/watch?v=yDzYesxqRdg&amp;list=PLT8pS9J_8jJqe2eAlpAng7KJY7uqK2NyR&amp;index=2&amp;t=38s</title>
      <dc:creator>Eze Lanza</dc:creator>
      <pubDate>Thu, 13 Mar 2025 18:50:11 +0000</pubDate>
      <link>https://dev.to/eze_lanza/quieres-aprender-sobre-agentes-en-espanol-he-creado-este-video-como-parte-de-una-serie-de-4i09</link>
      <guid>https://dev.to/eze_lanza/quieres-aprender-sobre-agentes-en-espanol-he-creado-este-video-como-parte-de-una-serie-de-4i09</guid>
      <description></description>
      <category>español</category>
      <category>tutorial</category>
    </item>
    <item>
      <title>S2E1 : Code &amp; Deploy: Data Contracts: Ensuring Reliable and Usable Data Products</title>
      <dc:creator>Eze Lanza</dc:creator>
      <pubDate>Thu, 20 Feb 2025 17:19:36 +0000</pubDate>
      <link>https://dev.to/eze_lanza/s2e1-code-deploy-data-contracts-ensuring-reliable-and-usable-data-products-2636</link>
      <guid>https://dev.to/eze_lanza/s2e1-code-deploy-data-contracts-ensuring-reliable-and-usable-data-products-2636</guid>
      <description>&lt;p&gt;Want to stop wasting time on data inconsistencies? Learn how the Open Data Contract Standard helps enforce data quality and structure, ensuring smooth data exchange between systems.&lt;/p&gt;

&lt;p&gt;Join Code &amp;amp; Deploy host Eze Lanza and his guests INNOQ's Jochen Christ and Chair of the TSC, Bitol, The Linux Foundation's Jean-Georges Perrin as they walk you through how to craft a data contract using the Open Data Contract Standard (ODCS) and how to use the open source Data Contract CLI tool to test the data contract and generate code.&lt;/p&gt;

&lt;p&gt;Project link Open Data Contract Standard (ODCS):&lt;a href="https://github.com/bitol-io/open-data-contract-standard/blob/main/README.md" rel="noopener noreferrer"&gt;https://github.com/bitol-io/open-data-contract-standard/blob/main/README.md&lt;/a&gt;&lt;br&gt;
Book : Implementing Data Mesh: Design, Build, and Implement Data Contracts, Data Products, and Data Mesh (&lt;a href="https://www.amazon.com/Implementing-Data-Mesh-Implement-Contracts-ebook/dp/B0DGYRBCZ2?ref_=ast_author_mpb" rel="noopener noreferrer"&gt;https://www.amazon.com/Implementing-Data-Mesh-Implement-Contracts-ebook/dp/B0DGYRBCZ2?ref_=ast_author_mpb&lt;/a&gt;)&lt;/p&gt;

</description>
      <category>ai</category>
      <category>developer</category>
      <category>livecoding</category>
    </item>
    <item>
      <title>https://www.linkedin.com/events/code-deploy-datacontracts-ensur7293365820594823168/theater/</title>
      <dc:creator>Eze Lanza</dc:creator>
      <pubDate>Wed, 19 Feb 2025 16:31:10 +0000</pubDate>
      <link>https://dev.to/eze_lanza/httpswwwlinkedincomeventscode-deploy-datacontracts-ensur7293365820594823168theater-4ekk</link>
      <guid>https://dev.to/eze_lanza/httpswwwlinkedincomeventscode-deploy-datacontracts-ensur7293365820594823168theater-4ekk</guid>
      <description></description>
      <category>career</category>
      <category>webdev</category>
    </item>
    <item>
      <title>FEB 19 11am EST! Code &amp; Deploy: Data Contracts: Ensuring Reliable and Usable Data Products</title>
      <dc:creator>Eze Lanza</dc:creator>
      <pubDate>Wed, 19 Feb 2025 04:12:45 +0000</pubDate>
      <link>https://dev.to/eze_lanza/feb-19-11am-est-code-deploy-data-contracts-ensuring-reliable-and-usable-data-products-19ec</link>
      <guid>https://dev.to/eze_lanza/feb-19-11am-est-code-deploy-data-contracts-ensuring-reliable-and-usable-data-products-19ec</guid>
      <description>&lt;p&gt;Are you struggling with inconsistent or unclear data definitions? Data contracts let you define clear, standardized expectations for your datasets. &lt;/p&gt;

&lt;p&gt;Join us tomorrow for our live stream about Data contracts!&lt;/p&gt;

&lt;p&gt;Sign in &lt;a href="https://www.linkedin.com/events/code-deploy-datacontracts-ensur7293365820594823168/comments/" rel="noopener noreferrer"&gt;https://www.linkedin.com/events/code-deploy-datacontracts-ensur7293365820594823168/comments/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Happy coding!&lt;/p&gt;

</description>
      <category>ai</category>
      <category>llm</category>
      <category>deepseek</category>
    </item>
    <item>
      <title>Top 4 Takeaways from DeepSeek-R1</title>
      <dc:creator>Eze Lanza</dc:creator>
      <pubDate>Thu, 06 Feb 2025 03:20:55 +0000</pubDate>
      <link>https://dev.to/eze_lanza/top-4-takeaways-from-deepseek-r1-23h5</link>
      <guid>https://dev.to/eze_lanza/top-4-takeaways-from-deepseek-r1-23h5</guid>
      <description>&lt;p&gt;Ok, it's been a week or two since the DeepSeek fever shook the AI world again (as it often happens when a new model appears). The initial reaction is often to figure out how to get hands-on experience with it, though running it locally is only feasible with the distilled versions unless you opt for alternative methods.&lt;br&gt;
What makes it particularly interesting is its ability to transfer reasoning capabilities to Small Language Models (SLMs) via distillation—from the massive 647B R1 down to 8B models using architectures like Llama and Qwen.&lt;/p&gt;

&lt;p&gt;After testing these distilled models, which are still &lt;strong&gt;reasoning models&lt;/strong&gt; at their core, we can observe they behave differently from traditional LLMs.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;A reasoning model has a totally different goal&lt;/strong&gt;. It has to evaluate all the options available to be sure the answer it gives is correct. &lt;/p&gt;

&lt;p&gt;What does this mean?&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;✅ Accuracy over speed: Reasoning models analyze multiple possibilities before generating responses, leading to longer processing times—but also ideally, more precise outputs.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;✅ Local testing: You can quickly prototype with the distilled models by downloading Ollama and running them locally. This also helps on scale deployments, on-prem and using APIs.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;❌ Open Source? : Not quite. Despite the trend toward openness, DeepSeek R1 itself is not open-source. (More on that in &lt;a href="https://www.linkedin.com/posts/jplorre_sam-altman-says-openai-is-on-the-wrong-side-activity-7292517736486715393-GEtP?utm_source=share&amp;amp;utm_medium=member_desktop&amp;amp;rcm=ACoAAAOrsgkBsW-GvMvp3mB3RlKCTcqeum_fJBc" rel="noopener noreferrer"&gt;this post&lt;/a&gt;).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;🚀 Reasoning in SLMs: The most exciting takeaway: SLMs with reasoning capabilities could reshape how we think about efficient, intelligent AI at small scales.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Did you try it?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Check the &lt;a href="https://community.intel.com/t5/Blogs/Tech-Innovation/Artificial-Intelligence-AI/Does-DeepSeek-Solve-the-Small-Scale-Model-Performance-Puzzle/post/1663390" rel="noopener noreferrer"&gt;full article&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Happy coding!&lt;/p&gt;

</description>
      <category>ai</category>
      <category>llm</category>
      <category>deepseek</category>
      <category>opensource</category>
    </item>
    <item>
      <title>Folks, we would love to hear from you!</title>
      <dc:creator>Eze Lanza</dc:creator>
      <pubDate>Wed, 18 Dec 2024 16:03:04 +0000</pubDate>
      <link>https://dev.to/eze_lanza/folks-we-would-love-to-hear-from-you-1k66</link>
      <guid>https://dev.to/eze_lanza/folks-we-would-love-to-hear-from-you-1k66</guid>
      <description>&lt;p&gt;Intel is conducting an Open Source Community Survey to better&lt;br&gt;
understand the needs and challenges of developers like you. With just fiveminutes of your time, you can help us create a more supportive and effective&lt;br&gt;
ecosystem. &lt;/p&gt;

&lt;p&gt;Complete the survey and enter to win one of four $250 cash gift&lt;br&gt;
cards! 🎁 &lt;/p&gt;

&lt;p&gt;Take the Survey: &lt;a href="https://lnkd.in/g3w26ia7" rel="noopener noreferrer"&gt;https://lnkd.in/g3w26ia7&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  OpenSource #Intel
&lt;/h1&gt;

</description>
      <category>survey</category>
      <category>developers</category>
      <category>ai</category>
      <category>opensource</category>
    </item>
    <item>
      <title>S1E3 : Code &amp; Deploy: Craft a Very Demure, Very Mindful Skincare Routine With GenA</title>
      <dc:creator>Eze Lanza</dc:creator>
      <pubDate>Thu, 26 Sep 2024 14:39:01 +0000</pubDate>
      <link>https://dev.to/eze_lanza/s1e3-code-deploy-craft-a-very-demure-very-mindful-skincare-routine-with-gena-33ll</link>
      <guid>https://dev.to/eze_lanza/s1e3-code-deploy-craft-a-very-demure-very-mindful-skincare-routine-with-gena-33ll</guid>
      <description>&lt;p&gt;Join Code &amp;amp; Deploy host Eze Lanza and special guest Aarushi Kansal, AI engineer at AutoGPT, as they show you how to build your own intelligent agent that can curate a personalized skincare routine which will leave you glowing. They’ll start with an introduction to the AutoGPT system, but you can choose to replicate the code in other frameworks or write it from scratch. Next, they’ll talk about retrieval augmented generation (RAG) and show you how to build a pipeline using a vector database—they’ll use Weaviate, but you can use whichever one you prefer. For this demo, Eze and Aarushi will be using Ollama [Mistral] so you can run the large language model (LLM) locally yourself.&lt;/p&gt;

&lt;p&gt;The main goal of the demo is to show you how to set up a reliable RAG-based system that minimizes hallucination, allows you to use your own data (without the overhead of fine-tuning), is controllable by you, and—bonus—leaves you with a plan for achieving great-looking skin.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>rag</category>
      <category>opensource</category>
      <category>programming</category>
    </item>
    <item>
      <title>Understanding Retrieval Augmented Generation (RAG)</title>
      <dc:creator>Eze Lanza</dc:creator>
      <pubDate>Wed, 28 Aug 2024 14:06:47 +0000</pubDate>
      <link>https://dev.to/eze_lanza/understanding-retrieval-augmented-generation-rag-f63</link>
      <guid>https://dev.to/eze_lanza/understanding-retrieval-augmented-generation-rag-f63</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;Learn what a RAG system is and how to deploy it using OPEA’s open source tools and frameworks&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;strong&gt;Originally posted on &lt;a href="https://medium.com/p/4d1d08f736b3" rel="noopener noreferrer"&gt;https://medium.com/p/4d1d08f736b3&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;By this point, most of us have used a large language model (LLM), like ChatGPT, to try to find quick answers to questions that rely on general knowledge and information. These questions range from the practical (What’s the best way to learn a new skill?) to the philosophical (What is the meaning of life?).&lt;br&gt;
&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1ze087ij6zbnq72r8dot.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1ze087ij6zbnq72r8dot.jpeg" alt="A screenshot listing the most common questions asked to an AI. The questions include topics like AI, homework help, the meaning of life, weather updates, creative writing, productivity, technology trends, and book or movie recommendations" width="720" height="620"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;But how do you get answers to questions that are personal? How much does your LLM know about you? Or your family?&lt;br&gt;
Let’s test ChatGPT and see how much it knows about my parents.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F635d47rjzly8n9t1eusc.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F635d47rjzly8n9t1eusc.jpeg" alt="A screenshot of a conversation where a user asks, ‘Do you know who is my mum?’ and the AI responds, explaining that it doesn’t have access to personal information about individuals." width="720" height="150"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;It’s understandable to feel frustrated when a model doesn’t recognize you, but it’s important to remember that these models don’t have much information about our personal lives. Unless you’re a celebrity or have your own Wikipedia page (as Tom Cruise has), the training dataset used for these models likely doesn’t include our information, which is why they can’t provide specific answers about us.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhz3vqpht674syhvzw85x.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhz3vqpht674syhvzw85x.jpeg" alt="A screenshot displays a question asking if the model knows who is Tom Cruise mum? the model answered Mary Lee Pfeiffer is his mom" width="720" height="143"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;So, how do we get our LLMs to know us better?&lt;/p&gt;

&lt;p&gt;That’s the million-dollar question facing enterprises looking to boost productivity with GenAI. They need models that provide context-based results. In this post, we’ll explain the basics of how retrieval augmented generation (RAG) improves your LLM’s responses and show you how to easily deploy your RAG-based model using a modular approach with the open source building blocks that are part of the new &lt;a href="https://opea.dev" rel="noopener noreferrer"&gt;Open Platform for Enterprise AI (OPEA)&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is RAG?
&lt;/h2&gt;

&lt;p&gt;We know that LLMs can greatly contribute to completing an extensive number of tasks, such as writing, learning, programming, translating, and more. However, the result we receive depends on what we ask the model, in other words, on how we meticulously build our prompts. For that reason, we spend too much time looking for the perfect prompt to get the answer we want; we’re starting to become experts in &lt;a href="https://en.wikipedia.org/wiki/Prompt_engineering" rel="noopener noreferrer"&gt;model prompting&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;Let’s return to the above question: “Who is my mum?” We know who our mum is, we have memories, and that information lives in our “mental” knowledge base, our brain.&lt;/p&gt;

&lt;p&gt;When building the prompt, we need to somehow provide it with memories of our mum and try to guide the model to use that information to creatively answer the question: Who is my mum? We’ll provide it with some of mum’s history and ask the model to take her past into account when answering the question.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbmh930heknhfvzl44vll.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbmh930heknhfvzl44vll.jpeg" alt="A screenshot displays a question asking for a creative response based on provided details: the user’s mother was born in the US, is 60, has strong Italian roots, and loves pizza. Below, a response discusses how her Italian heritage influences her personality, family traditions, and love for Italian cuisine, emphasizing her vibrant and lively nature." width="720" height="603"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As we can see, the model successfully gave us an answer that described my mum. Congratulations, we have used RAG!&lt;/p&gt;

&lt;p&gt;Let’s inspect what we did.&lt;/p&gt;

&lt;p&gt;Given the initial question, we tweaked the prompt to guide the model in how to use the information (context) we provided.&lt;/p&gt;

&lt;p&gt;We can think of the RAG process in three parts :&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Favqqeghcun9pg5sla7l3.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Favqqeghcun9pg5sla7l3.jpeg" alt="A screenshot shows a section divided into three parts. The top part labeled “Instruct” gives instructions asking to answer the question creatively, considering how roots influence the person’s behavior. The middle section, labeled “Context,” lists facts about the user’s mother: born in the US, age 60, strong Italian roots, and a love for pizza. The bottom part, labeled “Initial Question,” asks “Who is my mum?”" width="720" height="236"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Instruct: Guide the model. We have guided the model to use the information we provided (documents) to give us a creative answer and take into account my mum’s history. We used those instructions as an example; we could have used other guidance depending on the outcome we wanted to achieve. If we don’t want a creative answer, for example, this is the time to declare it.&lt;/li&gt;
&lt;li&gt;Context: Provide the context. In this example, we already knew the information about my mother since we retrieved that information from my memories, but in a real scenario, the challenge would be finding the relevant data in a knowledge base to feed the model so that it has the context needed to provide us with an accurate response, this process is called “retrieval.”&lt;/li&gt;
&lt;li&gt;Initial Question: The initial question we want answered.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Let’s explore how an enterprise can implement a real-life RAG example using open source tools and models. We’ll deploy it using the standardized frameworks and tools made available through OPEA, which was created to help streamline the implementation of enterprise AI.&lt;/p&gt;

&lt;h2&gt;
  
  
  Exploring the OPEA Architecture
&lt;/h2&gt;

&lt;p&gt;Here’s the architecture we used for the previous example:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpugmlds68qel02ix37nt.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpugmlds68qel02ix37nt.jpeg" alt="Image description" width="720" height="347"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;RAG can be understood as simply the steps mentioned above:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Initial Question&lt;/li&gt;
&lt;li&gt;Context&lt;/li&gt;
&lt;li&gt;Instruct&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;However, implementing the process in practice can be challenging because multiple components are needed: retrievers, embedding models, and a knowledge base, as shown in the image above. Let’s explore how those parts can work together.&lt;/p&gt;

&lt;p&gt;The key lies in providing the right context. You can compare the process to how our memories help us answer questions. For a company, this might mean drawing from a knowledge base of historical financial data or other relevant documents.&lt;/p&gt;

&lt;p&gt;For example, when a user asks a chatbot a question before the LLM can spit out an answer, the RAG application must first dive into a knowledge base and extract the most relevant information (the retrieval process). But even before the retrieval happens, an embedding model plays a crucial role in converting the data in the knowledge base into vector representations — meaningful numerical embeddings that capture the essence of the information. These embeddings will live in the knowledge base (vector database) and will allow the retriever to efficiently match the user’s query with the most relevant documents.&lt;/p&gt;

&lt;p&gt;Once the RAG application finds the relevant documents, it performs a rerank process to check the quality of the information and then re-orders the information based on relevance. It then builds a new prompt based on the refined context from the top-ranked documents and sends this prompt to the LLM, enabling the model to generate a high-quality, contextually informed response. Easy, right?&lt;/p&gt;

&lt;p&gt;As you can see, the RAG architecture isn’t about just one tool or one framework; it’s composed of multiple moving pieces making it difficult to pay attention to each component. When deploying a RAG system in our enterprise, we face multiple challenges, such as ensuring scalability, handling data security, and integrating with existing infrastructure.&lt;/p&gt;

&lt;p&gt;The Open Platform for Enterprise AI (OPEA) aims to solve those problems by treating each component in the RAG pipeline as a building block that is easily interchangeable. Say, for example, you’re using &lt;a href="https://huggingface.co/docs/transformers/en/model_doc/mistral" rel="noopener noreferrer"&gt;Mistral&lt;/a&gt;, but want to easily replace it with &lt;a href="https://huggingface.co/docs/transformers/en/model_doc/falcon" rel="noopener noreferrer"&gt;Falcon&lt;/a&gt;. Or, say you want to replace a vector database on the fly. You don’t want to have to rebuild the entire application. That would be a nightmare. OPEA makes deployment easier by providing robust tools and frameworks designed to streamline these processes and facilitate seamless integration.&lt;/p&gt;

&lt;p&gt;You can see this process in action by running the ChatQnA example: &lt;a href="https://github.com/opea-project/GenAIExamples/tree/main/ChatQnA" rel="noopener noreferrer"&gt;https://github.com/opea-project/GenAIExamples/tree/main/ChatQnA&lt;/a&gt;. There, you’ll find all the steps needed to create the building blocks for your RAG application on your server or your AIPC.&lt;/p&gt;

&lt;h2&gt;
  
  
  Call to Action
&lt;/h2&gt;

&lt;p&gt;We have shown you the basics of how RAG works and how to deploy a RAG pipeline using the OPEA framework. While the process is straightforward, deploying a RAG system at scale can introduce complexities. Here’s what you can do next:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Explore &lt;a href="https://github.com/opea-project/GenAIComps" rel="noopener noreferrer"&gt;GenAIComps&lt;/a&gt;: Gain insights into how generative AI components work together and how you can leverage them for real-world applications. OPEA provides detailed examples and documentation to guide your exploration.&lt;/li&gt;
&lt;li&gt;Explore &lt;a href="https://github.com/opea-project/GenAIExamples/tree/main/ChatQnA" rel="noopener noreferrer"&gt;RAG demo(ChatQnA)&lt;/a&gt;: Each part of a RAG system presents its own challenges, including ensuring scalability, handling data security, and integrating with existing infrastructure. OPEA, as an open source platform, offers tools and frameworks designed to address these issues and make the deployment process more efficient. Explore our demos to see how these solutions come together in practice.&lt;/li&gt;
&lt;li&gt;Explore &lt;a href="https://github.com/opea-project/GenAIExamples/tree/main" rel="noopener noreferrer"&gt;GenAI Examples&lt;/a&gt;: OPEA is not focused only on RAG; it is about generative AI as a whole. Multiple other demos, such as &lt;a href="https://github.com/opea-project/GenAIExamples/tree/main/VisualQnA" rel="noopener noreferrer"&gt;VisualQnA&lt;/a&gt;, showcase different GenAI capabilities. These examples demonstrate how OPEA can be leveraged across various tasks, expanding beyond RAG into other innovative GenAI applications.&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://github.com/opea-project/docs/blob/main/community/CONTRIBUTING.md" rel="noopener noreferrer"&gt;Contribute to the project&lt;/a&gt;! OPEA is built by a growing community of developers and AI professionals. Whether you’re interested in contributing code, improving documentation, or building new features, your involvement is key to our success.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Join us on the OPEA &lt;a href="https://github.com/opea-project" rel="noopener noreferrer"&gt;GitHub&lt;/a&gt; to start contributing or explore our issues list for ideas on where to start.&lt;/p&gt;

</description>
      <category>rag</category>
      <category>llm</category>
      <category>ai</category>
    </item>
    <item>
      <title>S1E2: Code &amp; Deploy: Build Your First Gen AI Agent with Haystack</title>
      <dc:creator>Eze Lanza</dc:creator>
      <pubDate>Mon, 26 Aug 2024 19:22:13 +0000</pubDate>
      <link>https://dev.to/eze_lanza/s1e2-code-deploy-build-your-first-gen-ai-agent-with-haystack-1c4h</link>
      <guid>https://dev.to/eze_lanza/s1e2-code-deploy-build-your-first-gen-ai-agent-with-haystack-1c4h</guid>
      <description>&lt;p&gt;Follow along with Intel Open Source AI Evangelist Ezequiel Lanza and his guest Bilge Yücel, developer relations engineer at deepset, as they show you how to build an intelligent GenAI agent using Haystack, an open source Python framework.&lt;/p&gt;

&lt;p&gt;In this step-by-step live demo, you'll learn how to build a customer support chatbot that can handle a wide range of user queries in real-time.&lt;/p&gt;

&lt;p&gt;Don't worry if you're new to Haystack! We'll start with a brief introduction to the framework, covering its core features and how it can be used to build robust and flexible AI applications. Following the introduction, we'll walk you through the process of creating a retrieval-augmented generation (RAG) pipeline using Haystack and the open source Mistral-7B-Instruct-v0.3 model, which is licensed under Apache-2.&lt;/p&gt;

&lt;p&gt;By integrating custom components, such as a Weather API Tool, you'll see how the chatbot handles specific requests, like checking the weather forecast, enhancing its functionality and providing a seamless user experience. This use case highlights the practical application of building adaptable AI agents with Haystack that can be deployed across various industries to improve customer interaction and satisfaction.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Haystack Basics + Web QA: &lt;a href="https://lnkd.in/dA_FU5Hg" rel="noopener noreferrer"&gt;https://lnkd.in/dA_FU5Hg&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;Build a Custom AI Agent with Haystack: &lt;a href="https://lnkd.in/dhcCuEUJ" rel="noopener noreferrer"&gt;https://lnkd.in/dhcCuEUJ&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>coding</category>
      <category>ai</category>
      <category>rag</category>
    </item>
    <item>
      <title>S1E1: Code &amp; Deploy: Build and Deploy an ML Binary Classifier</title>
      <dc:creator>Eze Lanza</dc:creator>
      <pubDate>Mon, 12 Aug 2024 23:05:59 +0000</pubDate>
      <link>https://dev.to/eze_lanza/code-deploy-build-and-deploy-an-ml-binary-classifier-4ho9</link>
      <guid>https://dev.to/eze_lanza/code-deploy-build-and-deploy-an-ml-binary-classifier-4ho9</guid>
      <description>&lt;p&gt;Join me with my guest William Arias, developer advocate at GitLab, as they show you step-by-step how to code and deploy a binary classifier using machine learning with open source tools. You’ll first prototype a solution in a Jupyter Notebook, then set up a basic CI pipeline to automate a machine learning app experimentation with MLFlow. Next, you’ll refactor the application (a bit) and create an API endpoint and simple UI for it to test manually (vibe testing). Before you’re done, you’ll learn how to combine everything and deploy the application to a runtime in the cloud using CI/CD principles. &lt;/p&gt;

&lt;p&gt;In this session, you’ll learn how to:&lt;/p&gt;

&lt;p&gt;• Prototype a binary classifier using machine learning in a Jupyter Notebook.&lt;/p&gt;

&lt;p&gt;• Set up a basic continuous integration (CI) pipeline with MLFlow for app experimentation.&lt;/p&gt;

&lt;p&gt;• Refactor the application to create an API endpoint and a simple user interface.&lt;/p&gt;

&lt;p&gt;• Deploy the application to a cloud runtime using continuous deployment (CD) principles.&lt;/p&gt;

&lt;p&gt;Links to repo&lt;br&gt;
 &lt;a href="https://gitlab.com/gitlab-da/use-cases/devsecops-platform/deep-learning/meowsky-classifier/-/blob/main/classifier/modeling/train.py" rel="noopener noreferrer"&gt;https://gitlab.com/gitlab-da/use-cases/devsecops-platform/deep-learning/meowsky-classifier/-/blob/main/classifier/modeling/train.py&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://gitlab.com/gitlab-da/use-cases/devsecops-platform/deep-learning/meowsky-classifier/-/blob/main/classifier/modeling/train.py" rel="noopener noreferrer"&gt;https://gitlab.com/gitlab-da/use-cases/devsecops-platform/deep-learning/meowsky-classifier/-/blob/main/classifier/modeling/train.py&lt;/a&gt;&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>ai</category>
      <category>coding</category>
    </item>
  </channel>
</rss>
