<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Farooq M</title>
    <description>The latest articles on DEV Community by Farooq M (@67d9a7007).</description>
    <link>https://dev.to/67d9a7007</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3844318%2F3ff51149-b2aa-4024-b2e1-13653ac51c5c.png</url>
      <title>DEV Community: Farooq M</title>
      <link>https://dev.to/67d9a7007</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/67d9a7007"/>
    <language>en</language>
    <item>
      <title>AIGoat - AI Security Playground to Attack and Defend LLMs. All Running Locally</title>
      <dc:creator>Farooq M</dc:creator>
      <pubDate>Thu, 26 Mar 2026 09:45:05 +0000</pubDate>
      <link>https://dev.to/67d9a7007/aigoat-we-built-an-ai-security-playground-to-attack-and-defend-llms-all-running-locally-555o</link>
      <guid>https://dev.to/67d9a7007/aigoat-we-built-an-ai-security-playground-to-attack-and-defend-llms-all-running-locally-555o</guid>
      <description>&lt;p&gt;We built an AI/LLM security playground - AI Goat where anyone from developers to security engineers can run a real AI application locally and start breaking it within minutes.&lt;/p&gt;

&lt;p&gt;No cloud setup. No API keys. No complex environment.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Just one command.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Once it’s running, you can:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;attack the system
&lt;/li&gt;
&lt;li&gt;exploit real vulnerabilities
&lt;/li&gt;
&lt;li&gt;switch between defense levels to see what actually works
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;All within the same application.&lt;/p&gt;

&lt;p&gt;This is what &lt;strong&gt;&lt;a href="https://github.com/AISecurityConsortium/AIGoat" rel="noopener noreferrer"&gt;AIGoat&lt;/a&gt;&lt;/strong&gt; is designed for.&lt;/p&gt;

&lt;p&gt;Getting Started Guide: &lt;a href="https://aigoat.co.in/blog/getting-started-with-aigoat/" rel="noopener noreferrer"&gt;https://aigoat.co.in/blog/getting-started-with-aigoat/&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;Most AI applications today are one prompt away from doing something they were never designed to do.&lt;/p&gt;

&lt;p&gt;And the scary part?&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Most teams don’t realize it.&lt;/strong&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  A Real Attack: Supply Chain Backdoor
&lt;/h2&gt;

&lt;p&gt;One of the most overlooked risks in AI systems is the &lt;strong&gt;supply chain&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;In AIGoat, we simulate this using a malicious model configuration.&lt;/p&gt;

&lt;p&gt;Here’s what happens:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;A model is shared publicly
&lt;/li&gt;
&lt;li&gt;It looks legitimate
&lt;/li&gt;
&lt;li&gt;It contains hidden behavioral triggers
&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;When integrated into an application:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;A specific phrase triggers data exfiltration
&lt;/li&gt;
&lt;li&gt;Another exposes internal prompts
&lt;/li&gt;
&lt;li&gt;Another manipulates business logic
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This is not traditional malware.&lt;/p&gt;

&lt;p&gt;This is &lt;strong&gt;behavioral compromise inside the AI itself&lt;/strong&gt;.&lt;/p&gt;




&lt;h2&gt;
  
  
  This Is Not Hypothetical
&lt;/h2&gt;

&lt;p&gt;This is exactly what we demonstrate in &lt;strong&gt;AIGoat&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;AIGoat is an open-source AI security playground where you can &lt;strong&gt;exploit real vulnerabilities in LLM-powered applications&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Not simulations.&lt;br&gt;&lt;br&gt;
Not slides.&lt;br&gt;&lt;br&gt;
A real, intentionally vulnerable AI system.&lt;/p&gt;


&lt;h2&gt;
  
  
  What Makes This Different?
&lt;/h2&gt;

&lt;p&gt;Most AI security discussions stay at:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;theory
&lt;/li&gt;
&lt;li&gt;best practices
&lt;/li&gt;
&lt;li&gt;high-level risks
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;AIGoat flips that model.&lt;/p&gt;

&lt;p&gt;You don’t just read about vulnerabilities.&lt;/p&gt;

&lt;p&gt;You exploit them yourself&lt;br&gt;&lt;br&gt;
You see how they break&lt;br&gt;&lt;br&gt;
You try to defend them  &lt;/p&gt;


&lt;h2&gt;
  
  
  Covering the OWASP Top 10 for LLMs
&lt;/h2&gt;

&lt;p&gt;AIGoat is designed around the &lt;strong&gt;OWASP Top 10 for LLM Applications&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Instead of just listing them, we:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Turn each category into a hands-on lab
&lt;/li&gt;
&lt;li&gt;Provide real attack scenarios
&lt;/li&gt;
&lt;li&gt;Let you test defensive controls
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;You can explore all of them in one place.&lt;/p&gt;

&lt;p&gt;As far as we know, there are very few platforms that allow you to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Learn
&lt;/li&gt;
&lt;li&gt;Exploit
&lt;/li&gt;
&lt;li&gt;Defend
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;all major LLM vulnerabilities end-to-end&lt;/strong&gt;&lt;/p&gt;


&lt;h2&gt;
  
  
  What You Get Inside AIGoat
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;17 hands-on attack labs
&lt;/li&gt;
&lt;li&gt;9 CTF-style challenges (auto-graded)
&lt;/li&gt;
&lt;li&gt;Multiple defense levels
&lt;/li&gt;
&lt;li&gt;Fully local setup (no cloud, no API keys)
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;You’re interacting with a &lt;strong&gt;real AI-powered e-commerce application&lt;/strong&gt;, not a toy example.&lt;/p&gt;


&lt;h2&gt;
  
  
  Why This Matters
&lt;/h2&gt;

&lt;p&gt;AI is being integrated into:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;customer support
&lt;/li&gt;
&lt;li&gt;financial workflows
&lt;/li&gt;
&lt;li&gt;internal tools
&lt;/li&gt;
&lt;li&gt;decision systems
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;But the security mindset hasn’t caught up.&lt;/p&gt;

&lt;p&gt;We’re still treating AI like:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“just another API”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;It’s not.&lt;/p&gt;

&lt;p&gt;It’s a &lt;strong&gt;dynamic, behavior-driven system&lt;/strong&gt; that can be manipulated in ways traditional systems cannot.&lt;/p&gt;


&lt;h2&gt;
  
  
  Try It Yourself
&lt;/h2&gt;

&lt;p&gt;Getting Started Guide: &lt;a href="https://aigoat.co.in/blog/getting-started-with-aigoat/" rel="noopener noreferrer"&gt;https://aigoat.co.in/blog/getting-started-with-aigoat/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;If you're:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;building with LLMs
&lt;/li&gt;
&lt;li&gt;securing AI systems
&lt;/li&gt;
&lt;li&gt;learning red teaming
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Spend 30 minutes with AIGoat.&lt;/p&gt;

&lt;p&gt;It will change how you think about AI security.&lt;/p&gt;


&lt;h2&gt;
  
  
  Get Started
&lt;/h2&gt;

&lt;p&gt;Website: &lt;a href="https://aigoat.co.in" rel="noopener noreferrer"&gt;https://aigoat.co.in&lt;/a&gt;&lt;br&gt;&lt;br&gt;
GitHub: &lt;a href="https://github.com/AISecurityConsortium/AIGoat" rel="noopener noreferrer"&gt;https://github.com/AISecurityConsortium/AIGoat&lt;/a&gt;  &lt;/p&gt;


&lt;h2&gt;
  
  
  Final Thought
&lt;/h2&gt;

&lt;p&gt;We don’t need more checklists for AI security.&lt;/p&gt;

&lt;p&gt;We need more people who have actually broken AI systems and understand how they fail.&lt;/p&gt;

&lt;p&gt;That’s what AIGoat is built for.&lt;/p&gt;


&lt;h2&gt;
  
  
  Feedback Welcome
&lt;/h2&gt;

&lt;p&gt;If you try it:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;break things
&lt;/li&gt;
&lt;li&gt;explore the labs
&lt;/li&gt;
&lt;li&gt;share feedback
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;We’re actively evolving the platform.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://github.com/AISecurityConsortium/AIGoat/stargazers" class="crayons-btn crayons-btn--primary" rel="noopener noreferrer"&gt;Star the repo. A GitHub star costs nothing and means everything. It helps other security practitioners discover the project.&lt;/a&gt;
&lt;/p&gt;




</description>
      <category>security</category>
      <category>ai</category>
      <category>llm</category>
      <category>rag</category>
    </item>
  </channel>
</rss>
