<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Anna Morelia</title>
    <description>The latest articles on DEV Community by Anna Morelia (@anna_morelia).</description>
    <link>https://dev.to/anna_morelia</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F2918312%2Fdd4b1780-d468-4193-aad1-cb1629f305ca.jpg</url>
      <title>DEV Community: Anna Morelia</title>
      <link>https://dev.to/anna_morelia</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/anna_morelia"/>
    <language>en</language>
    <item>
      <title>Let's Make It Easy M-C-Peasy</title>
      <dc:creator>Anna Morelia</dc:creator>
      <pubDate>Tue, 09 Sep 2025 17:44:14 +0000</pubDate>
      <link>https://dev.to/stacklok/lets-make-it-easy-m-c-peasy-3p0m</link>
      <guid>https://dev.to/stacklok/lets-make-it-easy-m-c-peasy-3p0m</guid>
      <description>&lt;h2&gt;
  
  
  🎥 Kicking off &lt;em&gt;Easy M-C-Peasy&lt;/em&gt;: A new video series on MCP
&lt;/h2&gt;

&lt;p&gt;The &lt;strong&gt;Model Context Protocol (MCP)&lt;/strong&gt; is starting to pop up everywhere if you’re building with agents or LLMs. There’s a lot of people writing about MCP, but if you’re new to it, it can feel a little overwhelming to piece together how it works, who to trust, and why you should care.  &lt;/p&gt;

&lt;p&gt;That’s why we're starting a new video series called &lt;strong&gt;Easy M-C-Peasy&lt;/strong&gt; with short, practical explainers of MCP basics, architecture, and more. &lt;/p&gt;

&lt;p&gt;In this first episode, we take on the foundational question:&lt;br&gt;&lt;br&gt;
&lt;strong&gt;“What is MCP and why does it matter?”&lt;/strong&gt;  &lt;/p&gt;

&lt;p&gt;&lt;em&gt;  &lt;iframe src="https://www.youtube.com/embed/9HAH-y_HbKk"&gt;
  &lt;/iframe&gt;
&lt;/em&gt;  &lt;/p&gt;

&lt;h2&gt;
  
  
  Why Stacklok cares
&lt;/h2&gt;

&lt;p&gt;At Stacklok, we’re working on &lt;a href="https://toolhive.dev" rel="noopener noreferrer"&gt;ToolHive&lt;/a&gt; — an open source project to make it easier (and safer) to connect AI agents with external tools. MCP is a big piece of that story, because it’s becoming the common language for how LLMs and agents plug into the world.  &lt;/p&gt;

&lt;p&gt;We don’t want MCP to feel like an insider-only protocol. The more people who understand it, the stronger the ecosystem gets — and the better we can all build. That’s why we’re investing time in making MCP approachable with this series.  &lt;/p&gt;

&lt;h2&gt;
  
  
  Full Playlist Available
&lt;/h2&gt;

&lt;p&gt;Dive into the entire &lt;a href="https://youtube.com/playlist?list=PLYBL38zBWVIjvdL5eWRIjjVRR9VdcViE1&amp;amp;feature=shared" rel="noopener noreferrer"&gt;YouTube playlist&lt;/a&gt; to explore every video back-to-back.&lt;/p&gt;

&lt;h2&gt;
  
  
  Share what you're curious about
&lt;/h2&gt;

&lt;p&gt;💡If you’ve got questions — about registries, security, auth, or just “how do I even get started?” — drop them in the comments. We’ll pull from those for future episodes.  &lt;/p&gt;

</description>
      <category>mcp</category>
      <category>ai</category>
      <category>modelcontextprotocol</category>
      <category>toolhive</category>
    </item>
    <item>
      <title>Which LLMs Are (and Aren't) Ready for Secure Code?</title>
      <dc:creator>Anna Morelia</dc:creator>
      <pubDate>Tue, 22 Apr 2025 18:52:48 +0000</pubDate>
      <link>https://dev.to/stacklok/which-llms-are-and-arent-ready-for-secure-code-38ac</link>
      <guid>https://dev.to/stacklok/which-llms-are-and-arent-ready-for-secure-code-38ac</guid>
      <description>&lt;h2&gt;Using the LLM Security Leaderboard to Select Models for Safe and Sustainable Code&lt;/h2&gt;

&lt;p&gt;Most language model benchmarking and comparison is focused on speed and accuracy. But with AI code generation, language model choice affects the safety and sustainability of resulting code. While many popular AI code-generation approaches rely on frontier models from providers like OpenAI and Anthropic, small- and mid-sized open-source models have advanced significantly and address specific needs for speed, efficiency, privacy, security, and compliance. To ensure developers and enterprises make informed choices, we’ve launched the &lt;a href="https://huggingface.co/spaces/stacklok/llm-security-leaderboard#/" rel="noopener noreferrer"&gt;LLM Security Leaderboard on Hugging Face&lt;/a&gt; to evaluate open-source models across four (initial) security dimensions. We’re taking an open, community-driven approach to this evaluation, and encourage you to join us in refining this benchmark. &lt;/p&gt;

&lt;p&gt;You can read more about &lt;a href="https://dev.to/stacklok/announcing-the-llm-security-leaderboard-evaluating-ai-models-through-a-security-lens-379n"&gt;our criteria and methodology here&lt;/a&gt;. Below are our takeaways from this first wave of analysis:&lt;/p&gt;

&lt;h2&gt;Key Findings&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;All models struggle with Bad Package Detection&lt;/strong&gt;: Llama 3.2-3B led, but only correctly flagged ~29% of bad NPM and PyPI packages. Nearly all the models we evaluated detected less than 5% of bad packages, and several popular models detected 0%, they simply provided instructions on how to install the package, regardless of whether the package existed or included typos. These models put the responsibility for bad package detection squarely on the user. &lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;CVE Knowledge is Alarmingly Low&lt;/strong&gt;: Awareness of  Common Vulnerabilities and Exposures (CVEs) in dependencies is a basic requirement for secure code. Yet most models scored between 8% and 18% accuracy in this category. Qwen2.5-Coder-3B-Instruct was the leader, but still scored low at 18.25%. These results suggest that the depth and consistency of CVE knowledge needs to be significantly improved.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Insecure Code Recognition is a Mixed Bag&lt;/strong&gt;: Top models like Qwen2.5-Coder-32B-Instruct and microsoft/phi-4 successfully identified vulnerabilities in roughly half of the code snippets presented. Lower-performing models recognized vulnerabilities in fewer than a quarter of cases; the inconsistency underscores the need for more targeted training on secure coding practices.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Model Size != Security&lt;/strong&gt;: While larger models often perform better on general benchmarks, security-specific performance varied significantly. Smaller models like Llama-3.2-3B-Instruct and IBM's Granite 3.3-2B-Instruct punched above their weight, reinforcing that sheer model size is not decisive and that architecture, training methodologies, and datasets play crucial roles in security capabilities.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Newer != Better&lt;/strong&gt;: Newer models like Qwen2.4-Coder-32b (knowledge cutoff June 24, 2024) and Granite-3.3-2b-Instruct (knowledge cutoff April 24, 2024) have about the same or lower bad package and CVE detection capabilities as older models like Llama-3.2-3b-Instruct (knowledge cutoff March 23, 2023), suggesting that these newer models were not trained on the latest bad package and CVE knowledge.&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;What This Means for Developers and Researchers&lt;/h2&gt;

&lt;p&gt;These findings should guide how teams approach secure AI adoption for software development:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Select models thoughtfully, especially when using LLMs in security-sensitive codegen workflows.&lt;/li&gt;
&lt;li&gt;Prioritize secure prompting techniques - careless prompting can exacerbate vulnerabilities.&lt;/li&gt;
&lt;li&gt;Complement LLMs with security-aware tools, like Stacklok's open-source project &lt;a href="https://github.com/stacklok/codegate" rel="noopener noreferrer"&gt;CodeGate&lt;/a&gt;, to reinforce defenses.&lt;/li&gt;
&lt;li&gt;Augment LLMs with Retrieval-Augmented Generation (RAG), using knowledge from leading vulnerability datasets such as &lt;a href="https://nvd.nist.gov/" rel="noopener noreferrer"&gt;NVD&lt;/a&gt;, &lt;a href="https://osv.dev/" rel="noopener noreferrer"&gt;OSV&lt;/a&gt;, &lt;a href="https://www.insight.stacklok.com/" rel="noopener noreferrer"&gt;Stacklok Insight&lt;/a&gt;, etc.&lt;/li&gt;
&lt;li&gt;Push for better fine-tuning and training on security datasets across the community.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;Get Involved&lt;/h2&gt;

&lt;p&gt;This is just the beginning. The &lt;a href="https://huggingface.co/spaces/stacklok/llm-security-leaderboard#/" rel="noopener noreferrer"&gt;LLM Security Leaderboard is live at Hugging Face&lt;/a&gt;, and we're inviting the community to submit models, suggest new evaluation methods, and contribute to a stronger, safer AI ecosystem.&lt;/p&gt;

&lt;p&gt;Explore the leaderboard. Submit your models. Join the conversation.&lt;/p&gt;

&lt;p&gt;Let's build a future where AI coding is safe and secure.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>security</category>
      <category>huggingface</category>
      <category>opensource</category>
    </item>
  </channel>
</rss>
