<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Karan Vyas</title>
    <description>The latest articles on DEV Community by Karan Vyas (@karan_vyas).</description>
    <link>https://dev.to/karan_vyas</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1581139%2F40849288-e44e-439d-84e8-ac45ec3a7f8f.jpg</url>
      <title>DEV Community: Karan Vyas</title>
      <link>https://dev.to/karan_vyas</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/karan_vyas"/>
    <language>en</language>
    <item>
      <title>Please help, I am new to Agentic LLM</title>
      <dc:creator>Karan Vyas</dc:creator>
      <pubDate>Wed, 05 Jun 2024 17:41:09 +0000</pubDate>
      <link>https://dev.to/karan_vyas/please-help-i-am-new-to-agentic-llm-3af5</link>
      <guid>https://dev.to/karan_vyas/please-help-i-am-new-to-agentic-llm-3af5</guid>
      <description>&lt;p&gt;Could someone help me with the most popular agentic LLM frameworks such as AutoGPT, AutoGen, Crew AI, LangGraph, etc.? &lt;br&gt;
I'm interested in setting them up locally with the most efficient configuration available. Have there been recent updates that enable them to run on top models like Phi-3 or LLAMA3 via local AI, etc.? &lt;br&gt;
I'm particularly keen on options and comparisons, considering that I operate on a machine with three RTX 6000 GPUs (Ada). &lt;br&gt;
What would be the best option/setup for optimizing performance with this setup?&lt;/p&gt;

</description>
      <category>llm</category>
      <category>genai</category>
      <category>agents</category>
    </item>
  </channel>
</rss>
