<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Julien Simon</title>
    <description>The latest articles on DEV Community by Julien Simon (@juliensimon).</description>
    <link>https://dev.to/juliensimon</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F88510%2F02d26c9a-2784-4b54-8a95-c48526734f3f.jpg</url>
      <title>DEV Community: Julien Simon</title>
      <link>https://dev.to/juliensimon</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/juliensimon"/>
    <language>en</language>
    <item>
      <title>Judging AFM-4.5B with DeepSeek-R1 670B</title>
      <dc:creator>Julien Simon</dc:creator>
      <pubDate>Thu, 03 Jul 2025 17:40:49 +0000</pubDate>
      <link>https://dev.to/juliensimon/judging-afm-45b-with-deepseek-r1-670b-1mkm</link>
      <guid>https://dev.to/juliensimon/judging-afm-45b-with-deepseek-r1-670b-1mkm</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjua4a911jz8ieujj4y71.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjua4a911jz8ieujj4y71.png"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;In this fun demonstration, you can witness the impressive capabilities of Arcee AI’s AFM-4.5B-Preview, Arcee’s first foundation model, across diverse domains. The demo showcases the model tackling complex knowledge questions, creating sophisticated creative writing, and addressing specialized domain-specific applications in healthcare, finance, technology, and education.&lt;/p&gt;

&lt;p&gt;Each response is independently evaluated by DeepSeek-R1 670B, providing expert assessment of answer quality. Will this small preview model be up to the task? Hey, what do you think? 😃&lt;/p&gt;

&lt;p&gt;  &lt;iframe src="https://www.youtube.com/embed/08rJPMYlTGg"&gt;
  &lt;/iframe&gt;
&lt;/p&gt;

</description>
      <category>ai</category>
      <category>slm</category>
      <category>datascience</category>
      <category>llm</category>
    </item>
    <item>
      <title>3 production-ready models released by Arcee AI on Hugging Face</title>
      <dc:creator>Julien Simon</dc:creator>
      <pubDate>Thu, 03 Jul 2025 12:59:45 +0000</pubDate>
      <link>https://dev.to/juliensimon/3-production-ready-models-released-by-arcee-ai-on-hugging-face-2bgn</link>
      <guid>https://dev.to/juliensimon/3-production-ready-models-released-by-arcee-ai-on-hugging-face-2bgn</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffktae0b27n0plsh70vz8.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffktae0b27n0plsh70vz8.png"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;In this video, I introduce and demonstrate three production-grade models that Arcee AI recently opened and released on Hugging Face.&lt;/p&gt;

&lt;p&gt;Arcee-SuperNova-v1 (70B) is a merged model built from multiple advanced training approaches. At its core is a distilled version of Llama-3.1–405B-Instruct into Llama-3.1–70B-Instruct, using our DistillKit to preserve instruction-following strengths while reducing size.&lt;/p&gt;

&lt;p&gt;Virtuoso-Large (72B) is our most powerful and versatile general-purpose model, designed to excel at handling complex and varied tasks across domains. With state-of-the-art performance, it offers unparalleled capability for nuanced understanding, contextual adaptability, and high accuracy.&lt;/p&gt;

&lt;p&gt;Caller (32B) is a robust model engineered for seamless integrations and optimized for managing complex tool-based interactions and API function calls. Its strength lies in precise execution, intelligent orchestration, and effective communication between systems, making it indispensable for sophisticated automation pipelines.&lt;/p&gt;

&lt;p&gt;  &lt;iframe src="https://www.youtube.com/embed/bjrWdcspsBA"&gt;
  &lt;/iframe&gt;
&lt;/p&gt;

</description>
      <category>slm</category>
      <category>llm</category>
      <category>ai</category>
      <category>enterprise</category>
    </item>
    <item>
      <title>Homunculus 12B and GLM-4–32B-Base-32K: 2 new Arcee AI research-oriented models</title>
      <dc:creator>Julien Simon</dc:creator>
      <pubDate>Thu, 03 Jul 2025 11:33:08 +0000</pubDate>
      <link>https://dev.to/juliensimon/homunculus-12b-and-glm-4-32b-base-32k-2-new-arcee-ai-research-oriented-models-4o1e</link>
      <guid>https://dev.to/juliensimon/homunculus-12b-and-glm-4-32b-base-32k-2-new-arcee-ai-research-oriented-models-4o1e</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqjoooq2b5ksfnmkz3roy.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqjoooq2b5ksfnmkz3roy.png"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;In this new video, I introduce two new research-oriented models that Arcee AI recently released on Hugging Face.&lt;/p&gt;

&lt;p&gt;Homunculus is a 12 billion-parameter instruction model distilled from Qwen3–235B onto the Mistral AI Nemo backbone. It was purpose-built to preserve Qwen’s two-mode interaction style — /think (deliberate chain-of-thought) and /nothink (concise answers) — while running on a single consumer GPU, and even on CPU as demonstrated in the video.&lt;/p&gt;

&lt;p&gt;GLM-4–32B-Base-32K is an enhanced version of Tsinghua University’s THUDM’s GLM-4–32B-Base-0414, specifically engineered to offer robust performance over an extended context window. While the original model’s capabilities degraded after 8,192 tokens, this version maintains strong performance up to a 32,000-token context, making it ideal for tasks requiring long-context understanding and processing.&lt;/p&gt;

&lt;p&gt;  &lt;iframe src="https://www.youtube.com/embed/mhrcPviW-MU"&gt;
  &lt;/iframe&gt;
&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>slm</category>
      <category>huggingface</category>
      <category>llm</category>
    </item>
    <item>
      <title>Introducing AFM, the Arcee Foundation Model</title>
      <dc:creator>Julien Simon</dc:creator>
      <pubDate>Wed, 18 Jun 2025 17:25:50 +0000</pubDate>
      <link>https://dev.to/juliensimon/introducing-afm-the-arcee-foundation-model-2778</link>
      <guid>https://dev.to/juliensimon/introducing-afm-the-arcee-foundation-model-2778</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fb3pidqxpbt19msh1siew.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fb3pidqxpbt19msh1siew.png" width="800" height="427"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Today, we’re thrilled to unveil the &lt;a href="https://www.arcee.ai/blog/announcing-the-arcee-foundation-model-family" rel="noopener noreferrer"&gt;Arcee Foundation Model&lt;/a&gt;, a new family of GenAI models built from the ground up for enterprise reality.&lt;/p&gt;

&lt;p&gt;The first release — AFM-4.5B — is a 4.5-billion-parameter frontier model that delivers excellent accuracy, strict compliance, and very high cost-efficiency.&lt;/p&gt;

&lt;p&gt;In short: enterprise-grade intelligence that can run anywhere — on a smartphone, at the edge, or in the cloud.&lt;/p&gt;

&lt;p&gt;For a quick taste, you can test AFM-4.5B in our &lt;a href="https://afm.arcee.ai" rel="noopener noreferrer"&gt;playground&lt;/a&gt; and on &lt;a href="https://api.together.ai/models/arcee-ai/AFM-4.5B-Preview" rel="noopener noreferrer"&gt;Together.ai&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;For a deeper dive into the model’s training pipeline and benchmarks, details are available in our &lt;a href="https://www.arcee.ai/blog/deep-dive-afm-4-5b-the-first-arcee-foundational-model" rel="noopener noreferrer"&gt;technical blog post&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;PS: As we’ll now focus on our AFM foundation models, and because we love open-source, we’re opening up access to our previously closed-source language models. Details in the tech blog post.&lt;/p&gt;

</description>
      <category>enterprise</category>
      <category>llm</category>
      <category>ai</category>
      <category>slm</category>
    </item>
    <item>
      <title>Introducing AFM, the Arcee Foundation Model — June 18</title>
      <dc:creator>Julien Simon</dc:creator>
      <pubDate>Wed, 04 Jun 2025 09:45:34 +0000</pubDate>
      <link>https://dev.to/juliensimon/introducing-afm-the-arcee-foundation-model-june-18-244f</link>
      <guid>https://dev.to/juliensimon/introducing-afm-the-arcee-foundation-model-june-18-244f</guid>
      <description>&lt;h3&gt;
  
  
  Introducing AFM, the Arcee Foundation Model — June 18
&lt;/h3&gt;

&lt;p&gt;Join us live on June 18 for the launch of AFM, the &lt;a href="http://www.arcee.ai" rel="noopener noreferrer"&gt;Arcee&lt;/a&gt; Foundation Model.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi5uoghwsxb7xywsjydrw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi5uoghwsxb7xywsjydrw.png" width="800" height="427"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Over the last two years, &lt;a href="http://www.arcee.ai" rel="noopener noreferrer"&gt;Arcee&lt;/a&gt; has proven the value of its post-training stack (MergeKit, DistillKit, EvolKit, Spectrum). Every open-source model we released on Hugging Face took the top spot on the OpenLLM Leaderboard in its size category.&lt;/p&gt;

&lt;p&gt;Now, &lt;a href="http://www.arcee.ai" rel="noopener noreferrer"&gt;Arcee&lt;/a&gt; is moving into building foundation models, and AFM, our first release, is about to redefine SLM performance. It outperforms all models in its size range, surpassing those from all top model builders. It also outperforms much larger models.&lt;/p&gt;

&lt;p&gt;Join us on June 18, we’ll tell you all about it.&lt;/p&gt;

&lt;p&gt;⭐️⭐️⭐️ YouTube stream: &lt;a href="https://youtube.com/live/QQv5P7jsc_E" rel="noopener noreferrer"&gt;https://youtube.com/live/QQv5P7jsc_E&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;⭐️⭐️⭐️ LinkedIn stream: &lt;a href="https://www.linkedin.com/events/7335944792935161856/" rel="noopener noreferrer"&gt;https://www.linkedin.com/events/7335944792935161856/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;⭐️⭐️⭐️ X stream: &lt;a href="https://x.com/julsimon" rel="noopener noreferrer"&gt;https://x.com/julsimon&lt;/a&gt;&lt;/p&gt;

</description>
      <category>opensource</category>
      <category>ai</category>
      <category>enterprise</category>
      <category>datascience</category>
    </item>
    <item>
      <title>AI at the edge — June 10, 2025 — live from Cisco Live in San Diego, CA!</title>
      <dc:creator>Julien Simon</dc:creator>
      <pubDate>Thu, 29 May 2025 10:31:57 +0000</pubDate>
      <link>https://dev.to/juliensimon/ai-at-the-edge-june-10-2025-live-from-cisco-live-in-san-diego-ca-24c4</link>
      <guid>https://dev.to/juliensimon/ai-at-the-edge-june-10-2025-live-from-cisco-live-in-san-diego-ca-24c4</guid>
      <description>&lt;h3&gt;
  
  
  AI at the edge — June 10, 2025 — live from Cisco Live in San Diego, CA!
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffatng3m2efxwr16mergp.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffatng3m2efxwr16mergp.png"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.linkedin.com/events/aiattheedge-livefromciscolivein7333796911532949504" rel="noopener noreferrer"&gt;Join us&lt;/a&gt; on June 10 for a fun session on AI at the edge, live from the floor of &lt;a href="https://www.linkedin.com/company/cisco/" rel="noopener noreferrer"&gt;Cisco&lt;/a&gt; Live in San Diego, CA!&lt;/p&gt;

&lt;p&gt;We’ll demonstrate and discuss an SLM-powered retail assistant running on &lt;a href="http://intel.com" rel="noopener noreferrer"&gt;Intel&lt;/a&gt; Xeon CPUs in a &lt;a href="http://cisco.com" rel="noopener noreferrer"&gt;Cisco&lt;/a&gt; UCS server.&lt;/p&gt;

&lt;p&gt;Thanks to a chatbot interface powered by open-source small language models and real-time data analytics, store associates can interact naturally through voice or text, receiving immediate information about product availability from &lt;a href="http://www.chooch.com" rel="noopener noreferrer"&gt;Chooch&lt;/a&gt;’s inventory system or crowd density from &lt;a href="http://www.thewaittimes.com" rel="noopener noreferrer"&gt;WaitTime&lt;/a&gt;’s analytics platform.&lt;/p&gt;

&lt;p&gt;AI at the edge is real, and you’ll see it with your own eyes. &lt;a href="https://www.linkedin.com/events/7333841334245494785/about/" rel="noopener noreferrer"&gt;Sign up now&lt;/a&gt;!&lt;/p&gt;

&lt;p&gt;You can also watch us live on YouTube:&lt;/p&gt;

&lt;p&gt;  &lt;iframe src="https://www.youtube.com/embed/iNbL3kIjMTI"&gt;
  &lt;/iframe&gt;
&lt;/p&gt;

</description>
      <category>enterprise</category>
      <category>retail</category>
      <category>ai</category>
      <category>llm</category>
    </item>
    <item>
      <title>Enriching Inventory Data with Arcee Conductor</title>
      <dc:creator>Julien Simon</dc:creator>
      <pubDate>Mon, 12 May 2025 14:09:33 +0000</pubDate>
      <link>https://dev.to/juliensimon/enriching-inventory-data-with-arcee-conductor-2amm</link>
      <guid>https://dev.to/juliensimon/enriching-inventory-data-with-arcee-conductor-2amm</guid>
      <description>&lt;p&gt;&lt;em&gt;This post was initially published on the Arcee AI &lt;em&gt;[_blog&lt;/em&gt;](&lt;a href="https://www.arcee.ai/blog/enriching-inventory-data-with-arcee-conductor" rel="noopener noreferrer"&gt;https://www.arcee.ai/blog/enriching-inventory-data-with-arcee-conductor&lt;/a&gt;)&lt;/em&gt;._&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffbuyc1yvwmbrjwne2ad1.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffbuyc1yvwmbrjwne2ad1.gif" width="600" height="337"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;In the world of inventory management, the accuracy and richness of data are paramount to help users and customers quickly and easily locate the right item every time. In mission-critical domains like healthcare, these items are not commodities; they are critical tools that can mean the difference between life and death. Ensuring that every item in the inventory is accurately described, categorized, and up-to-date is not just a best practice — it’s a necessity.&lt;/p&gt;

&lt;p&gt;‍&lt;/p&gt;

&lt;h3&gt;
  
  
  Inventory Data is Often Hard to Understand
&lt;/h3&gt;

&lt;p&gt;Unfortunately, many inventory management systems, particularly legacy systems, suffer from incomplete and inconsistent data. They often use heavily abbreviated descriptions due to application and database constraints. As a result, descriptions are hard for users to understand. Here’s an example you could find in a hospital inventory management system&lt;/p&gt;

&lt;p&gt;"Item": "IV START KIT W/CHG SKIN PREP CENTRAL LINE"&lt;/p&gt;

&lt;p&gt;If you’re an experienced doctor or nurse, you may figure it out. However, junior staff and non-medical staff would certainly be confused. Abbreviated descriptions often lack the necessary detail to distinguish between similar items. This ambiguity can lead to errors, especially in critical environments like hospitals, where the wrong tool can have severe consequences.&lt;/p&gt;

&lt;p&gt;The lack of detail also makes it difficult to implement user-friendly features in IT applications and severely limits search functionalities. Users may have to sift through multiple results to find the exact item they need whichwastes valuable time and increases the risk of errors. Personalized recommendations are also difficult to implement without additional data.&lt;/p&gt;

&lt;h3&gt;
  
  
  Language Models Can Make Inventory Data Human-Readable
&lt;/h3&gt;

&lt;p&gt;Thanks to data enrichment, we can add detailed descriptions and additional fields that significantly improve the user experience of inventory systems. Here’s how the example above can be improved, with a human-readable description and information on applications and risks.&lt;/p&gt;

&lt;p&gt;"Item": "IV START KIT W/CHG SKIN PREP CENTRAL LINE"&lt;br&gt;&lt;br&gt;
"Description": "An IV start kit that includes a chlorhexidine skin prep solution, designed for easy insertion and secure maintenance of central lines."&lt;br&gt;&lt;br&gt;
"Applications": "Insertion of central venous catheters", "Preparing the insertion site for IV access", "Infection control during line placement"&lt;br&gt;&lt;br&gt;
"Risks": "Potential for skin irritation from chlorhexidine", "Risk of infection if aseptic technique is not followed", "Allergic reactions to components of the kit"&lt;/p&gt;

&lt;p&gt;As we know by now, language models excel at understanding complex data. With the right prompt, they can easily generate the rich data we need to build better user experiences in inventory systems.&lt;/p&gt;

&lt;p&gt;One way to get started with an inventory system project would be to pick the “best” large language model (LLM) available today, and it could certainly do an excellent job. However, LLMs are notoriously slow and expensive, making it difficult to scale and customize the data enrichment process. Hospitals routinely manage thousands, sometimes tens of thousands, of unique inventory items. If you take a look at other industries like construction, food production, and of course, e-commerce, with a hospital you could be looking at 10x that number and even more.&lt;/p&gt;

&lt;p&gt;This is where &lt;a href="https://www.arcee.ai/product/arcee-conductor" rel="noopener noreferrer"&gt;Arcee Conductor&lt;/a&gt; can help.&lt;/p&gt;

&lt;h3&gt;
  
  
  Arcee Conductor Picks the Best Model for Each Prompt
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://www.arcee.ai/product/arcee-conductor" rel="noopener noreferrer"&gt;Arcee Conductor&lt;/a&gt; is a powerful inference platform based on a collection of high-quality small and large language models (LLMs). The beauty and efficiency of Arcee Conductor lie in its ability to select in real-time the most appropriate model for each query, ensuring that the output is both high-quality and cost-effective.&lt;/p&gt;

&lt;p&gt;Simple queries will automatically go to small language models (SLMs), delivering faster and more cost-effective inference than with an LLM. Only the more complex queries will go to LLMs, and you’ll only encounter their slower generation and higher cost when it’s truly necessary.&lt;/p&gt;

&lt;p&gt;Let’s see Conductor in action.&lt;/p&gt;

&lt;h3&gt;
  
  
  Arcee Conductor Can Efficiently Enrich Inventory Data
&lt;/h3&gt;

&lt;p&gt;We built a small demonstration to show how to enrich hospital inventory data with &lt;a href="https://www.arcee.ai/product/arcee-conductor" rel="noopener noreferrer"&gt;Arcee Conductor&lt;/a&gt;. The step-by-step Jupyter notebook is available on &lt;a href="https://gitlab.com/juliensimon/arcee-demos/-/tree/main/data-enrichment" rel="noopener noreferrer"&gt;GitLab&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;First, we generate 100 item descriptions similar to what you would find in a hospital inventory system. The file containing these descriptions is also available on &lt;a href="https://gitlab.com/juliensimon/arcee-demos/-/tree/main/data-enrichment" rel="noopener noreferrer"&gt;GitLab&lt;/a&gt;. Here are a few examples.&lt;/p&gt;

&lt;p&gt;“MASK SURG 3PLY ELASTIC EAR LOOP PLEATED DISP BFE98% ASTM2”&lt;br&gt;&lt;br&gt;
“GLOVE EXAM NITR PWD-FREE SML NONSTER TXTRD FINGRTIP”&lt;br&gt;&lt;br&gt;
“SYRINGE 3ML LUER-LOK TIP STERILE LATEX-FREE DISP”&lt;/p&gt;

&lt;p&gt;Then, we send each item description to the &lt;a href="https://www.arcee.ai/product/arcee-conductor" rel="noopener noreferrer"&gt;Arcee Conductor&lt;/a&gt; API configured in auto mode, letting it select the best SLM or LLM for each query. As the API is compatible with the OpenAI API, we can send our queries with the popular OpenAI client.&lt;/p&gt;

&lt;p&gt;Based on the abbreviated item description, we ask the model to write:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;A human-readable description, in 1–2 sentences,&lt;/li&gt;
&lt;li&gt;A list of applications,&lt;/li&gt;
&lt;li&gt;A list of risks,&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Once the process is over, we have successfully enriched the data! The enriched file is available on &lt;a href="https://gitlab.com/juliensimon/arcee-demos/-/tree/main/data-enrichment" rel="noopener noreferrer"&gt;GitLab&lt;/a&gt;, and the examples above now look like this.&lt;/p&gt;

&lt;p&gt;"Item": "MASK SURG 3PLY ELASTIC EAR LOOP PLEATED DISP BFE98% ASTM2"&lt;br&gt;&lt;br&gt;
"Description": "A surgical mask made of three layers of material, featuring elastic ear loops and pleated design, with a bacterial filtration efficiency of 98% as tested by ASTM2 standards."&lt;br&gt;&lt;br&gt;
"Applications": ["Protection against airborne particles during surgical procedures", "Prevention of cross-contamination in healthcare settings", "Use in environments requiring high-level respiratory protection"]&lt;br&gt;&lt;br&gt;
"Risks": ["Potential skin irritation from prolonged wear", "Risk of decreased breathability if wet or soiled", "Inadequate protection if not worn correctly or if damaged"]&lt;/p&gt;

&lt;p&gt;"Item": "GLOVE EXAM NITR PWD-FREE SML NONSTER TXTRD FINGRTIP"&lt;br&gt;&lt;br&gt;
"Description": "A powder-free, textured, nitrile exam glove designed for sensitive skin, providing protection during medical examinations."&lt;br&gt;&lt;br&gt;
"Applications": ["Use in clinical settings for patient examinations", "Procedures requiring tactile sensitivity"]&lt;br&gt;&lt;br&gt;
"Risks": ["Risk of allergic reactions to nitrile", "Potential for tears if not handled carefully"]}&lt;/p&gt;

&lt;p&gt;"Item": "SYRINGE 3ML LUER-LOK TIP STERILE LATEX-FREE DISP"&lt;br&gt;&lt;br&gt;
"Description": "A sterile, latex-free syringe with a 3ml capacity and a Luer-Lok tip, designed for precise and secure liquid delivery."&lt;br&gt;&lt;br&gt;
"Applications": ["Injection of medications", "Aspiration of fluids", "Administration of vaccines"]&lt;br&gt;&lt;br&gt;
"Risks": ["Risk of contamination if not properly sterilized", "Potential for air embolism if not used correctly", "Risk of needle stick injury if not handled properly"]}&lt;/p&gt;

&lt;p&gt;You can see that it’s much simpler to understand what the items are. The additional high-quality data also makes it easier to build efficient search or recommendation features.&lt;/p&gt;

&lt;p&gt;Now, let’s look at how cost-efficient this process is.&lt;/p&gt;

&lt;h3&gt;
  
  
  You Can Save up to 75% with Arcee Conductor
&lt;/h3&gt;

&lt;p&gt;For our 100 examples, here’s the breakdown of models selected by &lt;a href="https://www.arcee.ai/product/arcee-conductor" rel="noopener noreferrer"&gt;Arcee Conductor&lt;/a&gt;:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Arcee Virtuoso-Large, our 32-billion parameter general-purpose model: 86 times&lt;/li&gt;
&lt;li&gt;GPT-4.1: 7 times&lt;/li&gt;
&lt;li&gt;Claude Sonnet 3.7: 6 times&lt;/li&gt;
&lt;li&gt;Arcee Virtuoso-Medium, our 14-billion parameter general-purpose model: once&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Conductor was able to get the job done with Arcee SLMs 87% of the time. Not only does this reduce processing time by about 50%, but it also translates into significant cost savings. With a total cost of $0.0507, &lt;a href="https://www.arcee.ai/product/arcee-conductor" rel="noopener noreferrer"&gt;Arcee Conductor&lt;/a&gt; is respectively 75% and 56% more cost-effective than using only Sonnet 3.7 ($0.201) or GPT-4.1 ($0.1157).&lt;/p&gt;

&lt;p&gt;You may wonder, how did Arcee SLMs perform? Did they produce high-quality data similar to what the LLMs would have generated? Of course, we recommend that you run your own evaluation, but here’s a side-by-side comparison that is representative of the performance of our SLMs for this demo. We used the same prompt for the two models on this item description: “MASK SURG 3PLY ELASTIC EAR LOOP PLEATED DISP BFE98% ASTM2”&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Description&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;Virtuoso-Large: &lt;em&gt;A disposable surgical mask with elastic ear loops, pleated design, and a bacterial filtration efficiency of 98%, meeting ASTM level 2 standards.&lt;/em&gt;
&lt;/li&gt;
&lt;li&gt;Sonnet 3.7: &lt;em&gt;A disposable three-ply surgical mask with elastic ear loops, pleated design, and 98% bacterial filtration efficiency that meets ASTM Level 2 standards.&lt;/em&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Use cases&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;Virtuoso-Large:
&lt;em&gt;“Protecting healthcare workers during medical procedures”&lt;br&gt;
“Preventing the spread of respiratory infections in clinical settings”&lt;br&gt;
“Providing a barrier against large droplets and splashes”&lt;/em&gt;
&lt;/li&gt;
&lt;li&gt;Sonnet 3.7:
&lt;em&gt;“General medical procedures”&lt;br&gt;
“Healthcare settings”&lt;br&gt;
“Patient care areas”&lt;br&gt;
“Protection against respiratory droplets”&lt;br&gt;
“Surgical environments”&lt;/em&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Risks&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;Virtuoso-Large:
&lt;em&gt;“Potential for reduced effectiveness if not properly fitted or worn”&lt;br&gt;
“Risk of contamination if reused or not disposed of properly”&lt;br&gt;
“May cause skin irritation or allergic reactions in some users.”&lt;/em&gt;
&lt;/li&gt;
&lt;li&gt;Sonnet 3.7:
&lt;em&gt;“Not sufficient for aerosol-generating procedures”&lt;br&gt;
“May not provide adequate protection against airborne pathogens”&lt;br&gt;
“Single-use only”&lt;br&gt;
“Improper fit may reduce effectiveness”&lt;br&gt;
“Does not create a complete seal around face”&lt;/em&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Cost&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;Virtuoso-Large: $0.00026685
&lt;/li&gt;
&lt;li&gt;Sonnet 3.7: $0.002595&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Time&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;Virtuoso-Large: 1.87 second
&lt;/li&gt;
&lt;li&gt;Sonnet 3.7: 3.2 seconds&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;‍As you can see, the results are extremely close, and it’s hard to justify the 10x higher cost of Sonnet and its 70% slower generation time.&lt;/p&gt;

&lt;h3&gt;
  
  
  Conclusion
&lt;/h3&gt;

&lt;p&gt;Using &lt;a href="https://www.arcee.ai/product/arcee-conductor" rel="noopener noreferrer"&gt;Arcee Conductor&lt;/a&gt; for data enrichment can transform your inventory management systems, and generally, any system that could benefit from better, richer data and metadata. The automatic model selection ensures that you are always using the most appropriate model, and the cost savings are very significant. We encourage you to try it!&lt;/p&gt;

&lt;p&gt;We’d love to hear from you and see how we can help. Don’t hesitate to contact &lt;a href="//mailto:sales@arcee.ai"&gt;sales@arcee.ai&lt;/a&gt; or submit a request to meet with our &lt;a href="https://www.arcee.ai/book-a-demo" rel="noopener noreferrer"&gt;book a demo&lt;/a&gt; form.&lt;/p&gt;

&lt;h3&gt;
  
  
  Resources
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Learn more about &lt;a href="https://www.arcee.ai/product/arcee-conductor" rel="noopener noreferrer"&gt;Arcee Conductor&lt;/a&gt;.&lt;/li&gt;
&lt;li&gt;
&lt;a href="http://conductor.arcee.ai" rel="noopener noreferrer"&gt;Sign up&lt;/a&gt; for Arcee Conductor and get $20 of free inference credits.&lt;/li&gt;
&lt;li&gt;Read the Arcee Conductor &lt;a href="https://docs.arcee.ai/arcee-conductor/introduction-to-arcee-conductor" rel="noopener noreferrer"&gt;documentation&lt;/a&gt;.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Grab the notebook and sample files on &lt;a href="https://gitlab.com/juliensimon/arcee-demos/-/tree/main/data-enrichment" rel="noopener noreferrer"&gt;GitLab&lt;/a&gt;.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>enterprise</category>
      <category>llm</category>
    </item>
    <item>
      <title>Announcing Arcee AI AnyMCP</title>
      <dc:creator>Julien Simon</dc:creator>
      <pubDate>Thu, 08 May 2025 16:42:53 +0000</pubDate>
      <link>https://dev.to/juliensimon/announcing-arcee-ai-anymcp-4nl9</link>
      <guid>https://dev.to/juliensimon/announcing-arcee-ai-anymcp-4nl9</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcunto8vo51js1f02fygg.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcunto8vo51js1f02fygg.png"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;We’re thrilled to announce the first release of *&lt;em&gt;Arcee AI *&lt;/em&gt; &lt;a href="https://mcp.arcee.ai" rel="noopener noreferrer"&gt;&lt;strong&gt;AnyMCP&lt;/strong&gt;&lt;/a&gt;!&lt;/p&gt;

&lt;p&gt;At Arcee, we believe in making powerful technology accessible and easy to use. That’s why we’ve developed Arcee AI &lt;a href="https://mcp.arcee.ai" rel="noopener noreferrer"&gt;AnyMCP&lt;/a&gt;, a groundbreaking platform that allows you to remotely deploy and manage thousands of MCP (Multi-Client Protocol) servers in seconds. Whether you’re a developer, a gamer, or a tech enthusiast, Arcee AI &lt;a href="https://mcp.arcee.ai" rel="noopener noreferrer"&gt;AnyMCP&lt;/a&gt; is designed to streamline your workflow and enhance your experience.&lt;/p&gt;

&lt;h4&gt;
  
  
  Why Arcee AI AnyMCP?
&lt;/h4&gt;

&lt;p&gt;In the words of the King, “🎶 a little less conversation, a little more action 🎶.” We couldn’t agree more. The world of MCP can be complex, but it doesn’t have to be. With Arcee AI &lt;a href="https://mcp.arcee.ai" rel="noopener noreferrer"&gt;AnyMCP&lt;/a&gt;, you can focus on what matters most — action and innovation&lt;/p&gt;

&lt;p&gt;Here are the key features:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;One-Click Deployment&lt;/strong&gt; : Deploy any MCP server with a single click and start using it immediately with any MCP-compatible client.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Seamless Integration&lt;/strong&gt; : Integrate seamlessly with your existing tools and workflows.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Remote Deployment &amp;amp; Management&lt;/strong&gt; : Deploy and manage thousands of MCP servers with just a few clicks.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Compatibility&lt;/strong&gt; : Use Arcee AI &lt;a href="https://mcp.arcee.ai" rel="noopener noreferrer"&gt;AnyMCP&lt;/a&gt; with Claude Desktop or any other MCP-compatible client.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Scalability&lt;/strong&gt; : Supports thousands of MCP servers. If your server isn’t listed, just request it!&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;100% Free&lt;/strong&gt; : Yes, you read that right! Arcee AI &lt;a href="https://mcp.arcee.ai" rel="noopener noreferrer"&gt;AnyMCP&lt;/a&gt; is completely free to use right now.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;  &lt;iframe src="https://www.youtube.com/embed/cAjG5xILUcY"&gt;
  &lt;/iframe&gt;
&lt;/p&gt;

&lt;h4&gt;
  
  
  Get Started Today!
&lt;/h4&gt;

&lt;p&gt;Ready to experience the power of Arcee AI &lt;a href="https://mcp.arcee.ai" rel="noopener noreferrer"&gt;AnyMCP&lt;/a&gt;? Head over to &lt;a href="https://mcp.arcee.ai/" rel="noopener noreferrer"&gt;https://mcp.arcee.ai/&lt;/a&gt;to get started.&lt;/p&gt;

&lt;p&gt;Let’s make the world of MCP more accessible and exciting together! 🤘&lt;/p&gt;

&lt;p&gt;We can’t wait to see what you create with Arcee AI AnyMCP. Try it now and share your thoughts with us. Your feedback is invaluable as we continue to improve and expand our platform.&lt;/p&gt;

&lt;p&gt;Happy deploying! 🚀&lt;/p&gt;

</description>
      <category>ai</category>
      <category>mcp</category>
      <category>llm</category>
    </item>
    <item>
      <title>Arcee AI webinar: routing your function calling and reasoning queries with Arcee Conductor</title>
      <dc:creator>Julien Simon</dc:creator>
      <pubDate>Wed, 07 May 2025 08:31:48 +0000</pubDate>
      <link>https://dev.to/juliensimon/arcee-ai-webinar-routing-your-function-calling-and-reasoning-queries-with-arcee-conductor-7n</link>
      <guid>https://dev.to/juliensimon/arcee-ai-webinar-routing-your-function-calling-and-reasoning-queries-with-arcee-conductor-7n</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffy4gkhvgnwsoze1qqm86.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffy4gkhvgnwsoze1qqm86.png"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;In this video, we show you how &lt;a href="https://www.arcee.ai/product/arcee-conductor" rel="noopener noreferrer"&gt;Arcee Conductor&lt;/a&gt; can now automatically route each function calling or reasoning query to the best SLM/LLM, efficiently delivering precise and cost-effective results for any task.&lt;/p&gt;

&lt;p&gt;Sign up for Conductor and get $20 of free inference credits!&lt;/p&gt;

&lt;p&gt;  &lt;iframe src="https://www.youtube.com/embed/JfyvMxMPfJY"&gt;
  &lt;/iframe&gt;
&lt;/p&gt;

</description>
      <category>ai</category>
      <category>enterprise</category>
      <category>datascience</category>
      <category>agentic</category>
    </item>
    <item>
      <title>Model routing with reasoning queries in Arcee Conductor</title>
      <dc:creator>Julien Simon</dc:creator>
      <pubDate>Tue, 06 May 2025 10:21:43 +0000</pubDate>
      <link>https://dev.to/juliensimon/model-routing-with-reasoning-queries-in-arcee-conductor-1i0m</link>
      <guid>https://dev.to/juliensimon/model-routing-with-reasoning-queries-in-arcee-conductor-1i0m</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffe2s9jsbzh52y3gkxdxz.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffe2s9jsbzh52y3gkxdxz.png"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.arcee.ai/product/arcee-conductor" rel="noopener noreferrer"&gt;Arcee Conductor&lt;/a&gt; now supports routing for reasoning models. You can automatically each one of your reasoning queries to the best and most effective SLM/LLM 😊&lt;/p&gt;

&lt;p&gt;  &lt;iframe src="https://www.youtube.com/embed/KTEgZW9GNeU"&gt;
  &lt;/iframe&gt;
&lt;/p&gt;

</description>
      <category>llm</category>
      <category>enterprise</category>
      <category>ai</category>
    </item>
    <item>
      <title>Arm podcast</title>
      <dc:creator>Julien Simon</dc:creator>
      <pubDate>Mon, 21 Apr 2025 18:18:23 +0000</pubDate>
      <link>https://dev.to/juliensimon/arm-podcast-5jd</link>
      <guid>https://dev.to/juliensimon/arm-podcast-5jd</guid>
      <description>&lt;p&gt;The podcast I recently recorded with &lt;a href="http://www.arm.com" rel="noopener noreferrer"&gt;Arm&lt;/a&gt; is now available on all major platforms. Tune in if you want to learn about the evolution of small language models, the significance of CPU-based AI inference, and what &lt;a href="http://www.arcee.ai" rel="noopener noreferrer"&gt;Arcee AI&lt;/a&gt; is doing in that space.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://podcasts.apple.com/gb/podcast/arm-viewpoints-small-language-models-big-ambitions/id1528983070?i=1000703527091" rel="noopener noreferrer"&gt;Arm Viewpoints: Small language models, big ambitions&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Model routing for function calling with Arcee Conductor</title>
      <dc:creator>Julien Simon</dc:creator>
      <pubDate>Thu, 10 Apr 2025 10:16:39 +0000</pubDate>
      <link>https://dev.to/juliensimon/model-routing-for-function-calling-with-arcee-conductor-5462</link>
      <guid>https://dev.to/juliensimon/model-routing-for-function-calling-with-arcee-conductor-5462</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq415c8zovz617nj9hye3.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq415c8zovz617nj9hye3.png"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;In this video, we show you how to run function calling with &lt;a href="https://www.arcee.ai/product/arcee-conductor" rel="noopener noreferrer"&gt;Arcee Conductor&lt;/a&gt;. This allows you to automatically invoke external APIs and tools, and use their output for text generation while making sure you use the best and most cost-effective SLM/LLM for each query. In this particular example, I run API calls on Yahoo Finance.&lt;/p&gt;

&lt;p&gt;  &lt;iframe src="https://www.youtube.com/embed/XAwi_8CoNE8"&gt;
  &lt;/iframe&gt;
&lt;/p&gt;

</description>
      <category>enterprise</category>
      <category>ai</category>
      <category>automation</category>
      <category>datascience</category>
    </item>
  </channel>
</rss>
