<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Dasha Maliugina</title>
    <description>The latest articles on DEV Community by Dasha Maliugina (@dasha_maliugina).</description>
    <link>https://dev.to/dasha_maliugina</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1724969%2Ff939e9a3-1fd2-4b40-86b6-21fec82cbf0b.jpg</url>
      <title>DEV Community: Dasha Maliugina</title>
      <link>https://dev.to/dasha_maliugina</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/dasha_maliugina"/>
    <language>en</language>
    <item>
      <title>💡 10 learnings on LLM evaluations</title>
      <dc:creator>Dasha Maliugina</dc:creator>
      <pubDate>Wed, 14 May 2025 19:29:44 +0000</pubDate>
      <link>https://dev.to/dasha_maliugina/10-learnings-on-llm-evaluations-1lm6</link>
      <guid>https://dev.to/dasha_maliugina/10-learnings-on-llm-evaluations-1lm6</guid>
      <description>&lt;p&gt;If you are building with LLMs, you already know this: evaluation is hard. And it's nothing like testing regular software. You can’t just throw in some unit tests and call it a day. &lt;/p&gt;

&lt;p&gt;Here are 10 key ideas on LLM evaluations, sourced from &lt;a href="https://www.evidentlyai.com/llm-evaluation-course-practice" rel="noopener noreferrer"&gt;LLM evaluation for builders: applied course&lt;/a&gt; (It’s a free and open course with YouTube videos and code examples).&lt;/p&gt;

&lt;h2&gt;
  
  
  1. LLM evaluation ≠ Benchmarking
&lt;/h2&gt;

&lt;p&gt;Evaluating LLM systems is not the same as evaluating LLM models. Public &lt;a href="https://www.evidentlyai.com/llm-guide/llm-benchmarks" rel="noopener noreferrer"&gt;LLM benchmarks&lt;/a&gt; test general model capabilities, like math, code, or reasoning. But if you're building an LLM application, you need to evaluate it on your specific use cases.&lt;/p&gt;

&lt;p&gt;For example, for a customer support chatbot, you need to test how well it grounds the answers on the company’s policies or handles tricky queries like comparing your product against competitors. It’s not just how it performs on MMLU.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F60l566je3rexl78nv4k6.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F60l566je3rexl78nv4k6.png" alt="LLM benchmarks vs LLM evals" width="800" height="800"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  2. Evaluation is a tool, not a task
&lt;/h2&gt;

&lt;p&gt;LLM evaluation isn’t just a checkbox. It is a tool that helps answer product questions and supports decision-making throughout the product lifecycle:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;During the experiments&lt;/strong&gt;, evals help you compare different prompts, models, and settings to determine what works best. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Before deployment&lt;/strong&gt;, you can run stress-testing and red-teaming to check how your system handles edge cases.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Once in production&lt;/strong&gt;, you need to monitor how well your product is doing in the wild.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;When you make changes&lt;/strong&gt;, you need to run regression tests before shipping the updates.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;So, to design evaluations well, you first need to figure out what you are looking to solve!&lt;/p&gt;

&lt;h2&gt;
  
  
  3. LLM evaluation ≠ Software testing
&lt;/h2&gt;

&lt;p&gt;Unlike traditional software, LLM systems are non-deterministic. That means that the same input can yield different outputs. Also, LLM products often solve open-ended tasks, like writing an email, that do not have a single correct answer. &lt;/p&gt;

&lt;p&gt;In addition, LLM systems bring a whole new set of risks. They can hallucinate and confidently make things up. Malicious users can attempt to jailbreak LLM apps and bypass their security measures. LLM app data leaks can result in exposing sensitive data.&lt;/p&gt;

&lt;p&gt;Testing for functionality is no longer enough. You must also evaluate the quality and safety of your LLM system’s responses.&lt;/p&gt;

&lt;h2&gt;
  
  
  4. Combine manual and automated evaluations
&lt;/h2&gt;

&lt;p&gt;You should always start with manual review to build intuition. Your goal is to understand what “good” means for your use case and spot patterns: are there any common failure modes or unexpected behavior? &lt;/p&gt;

&lt;p&gt;Once you know what you’re looking for, you can add automation. But the important thing is that automated LLM evals are here to scale human judgment, not replace it.&lt;/p&gt;

&lt;h2&gt;
  
  
  5. Use both reference-based and reference-free evals
&lt;/h2&gt;

&lt;p&gt;There are two main types of &lt;a href="https://www.evidentlyai.com/llm-guide/llm-evaluation-metrics" rel="noopener noreferrer"&gt;LLM evaluation methods&lt;/a&gt;:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Reference-based evals&lt;/strong&gt; compare your system’s outputs to expected – or “ground-truth” – answers, which is great for regression testing or experiments. You can use such methods as exact match, semantic similarity, BERTScore, and LLM-as-a-judge.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Reference-free evals&lt;/strong&gt; allow assessing specific qualities of the response, like helpfulness or tone. These are useful in open-ended scenarios and production monitoring. You can use text statistics, regular expressions, ML models, and LLM judges. &lt;/p&gt;

&lt;p&gt;As you’ve probably guessed, you’ll need both types!&lt;/p&gt;

&lt;h2&gt;
  
  
  6. Think in datasets, not unit tests
&lt;/h2&gt;

&lt;p&gt;Traditional testing is built around unit tests. With LLMs, it’s more useful to think in datasets. You need to test for a range of acceptable behaviors, so it’s not enough to run evaluations on a single example. &lt;/p&gt;

&lt;p&gt;You may need to create diverse test sets, including happy paths, edge cases, and adversarial examples. &lt;a href="https://www.evidentlyai.com/llm-guide/llm-test-dataset-synthetic-data" rel="noopener noreferrer"&gt;Good evaluation datasets&lt;/a&gt; reflect both how users interact with your app in the real world and where things can go wrong.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fb48lq70k48kyz3430z40.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fb48lq70k48kyz3430z40.png" alt="Edge case example" width="800" height="800"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  7. LLM-as-a-judge is a key evaluation method
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://www.evidentlyai.com/llm-guide/llm-as-a-judge" rel="noopener noreferrer"&gt;LLM-as-a-judge&lt;/a&gt; is a common technique to evaluate LLM-powered products. The idea is simple: you can use another LLM (or the same one!) to evaluate your system’s response with a custom evaluation prompt. For example, you can ask the judge whether your chatbot’s response is polite or aligns with the brand image. &lt;/p&gt;

&lt;p&gt;This approach is scalable and surprisingly effective. Just remember, LLM judges aren’t perfect. You’ll need to assess and tune them, and invest in designing the evaluation criteria.&lt;/p&gt;

&lt;h2&gt;
  
  
  8. Use custom criteria, not just generic metrics
&lt;/h2&gt;

&lt;p&gt;Since LLMs often solve very custom tasks, you must design quality criteria that map to your use case. You can’t just blindly use metrics like coherence or helpfulness without critically thinking about what actually matters.&lt;/p&gt;

&lt;p&gt;Instead, you must define what “good” means for your app. Then, customize your evaluation to your domain, users, and specific risks. For a legal assistant, you can check whether the answer cites the correct regulations. For a wellness chatbot, you may need to ensure it answers in a friendly manner and does not provide medical advice. &lt;/p&gt;

&lt;h2&gt;
  
  
  9. Start with analytics
&lt;/h2&gt;

&lt;p&gt;Evaluations are very analytical tasks.&lt;/p&gt;

&lt;p&gt;To run LLM evals, you first need the data. So log everything: capture all the inputs and outputs, record metadata like model version and prompt, and track user feedback if you have it. If your app doesn’t have real users yet, you can start with synthetic data and grow from there.&lt;/p&gt;

&lt;p&gt;You also need to manually analyze the outputs you get to determine your criteria and understand the failure modes you observe.&lt;/p&gt;

&lt;h2&gt;
  
  
  10. Evaluation is a moat
&lt;/h2&gt;

&lt;p&gt;Building a solid LLM evaluation system is an investment. But it is also a competitive advantage:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Rapid iteration.&lt;/strong&gt; Evals help speed up your AI product development cycle, ship updates stress-free, switch models easily, and debug issues efficiently. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Safe deployment.&lt;/strong&gt; Evaluations allow you to test how an LLM system handles edge cases to avoid liability risks and protect customers from harmful outputs.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Product quality at scale.&lt;/strong&gt; Finally, evals help ensure your LLM app works well and provides a good customer experience. &lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Your competitors can’t copy that — even if they use the same model for their LLM app!&lt;/p&gt;

&lt;h2&gt;
  
  
  🔥 Free course on LLM evaluations
&lt;/h2&gt;

&lt;p&gt;Learn how to create LLM judges, evaluate RAG systems, and run adversarial tests. The course is designed for AI/ML Engineers and those building real-world LLM apps; basic Python skills are required. And yes, it’s free! &lt;a href="https://www.evidentlyai.com/llm-evaluation-course-practice" rel="noopener noreferrer"&gt;Learn more and sign up&lt;/a&gt;.&lt;/p&gt;

</description>
      <category>llm</category>
      <category>ai</category>
      <category>genai</category>
      <category>machinelearning</category>
    </item>
    <item>
      <title>20 examples of LLM-powered applications in the real world</title>
      <dc:creator>Dasha Maliugina</dc:creator>
      <pubDate>Wed, 03 Jul 2024 17:13:55 +0000</pubDate>
      <link>https://dev.to/dasha_maliugina/20-examples-of-llm-powered-applications-in-the-real-world-p8c</link>
      <guid>https://dev.to/dasha_maliugina/20-examples-of-llm-powered-applications-in-the-real-world-p8c</guid>
      <description>&lt;p&gt;The recent advancements in LLMs improved their performance and made them more affordable – this unlocked multiple possibilities for companies to integrate LLMs into their products. Indeed, there have been a lot of impressive demos. But how do companies actually use LLMs in production? &lt;/p&gt;

&lt;p&gt;We put together and regularly update a &lt;a href="https://www.evidentlyai.com/ml-system-design"&gt;database&lt;/a&gt; of 450 use cases from 100+ companies that detail real-world applications and insights from ML and LLM system design. In this blog, we share 20 selected examples of LLM-powered products from various industries. &lt;/p&gt;

&lt;p&gt;The database is maintained by the team behind Evidently, an open-source tool for LLM and ML evaluation and observability. &lt;a href="https://github.com/evidentlyai/evidently"&gt;Give us a star on GitHub&lt;/a&gt; to support the project! &lt;/p&gt;

&lt;h2&gt;
  
  
  👷 &lt;a href="https://www.linkedin.com/blog/engineering/skills-graph/extracting-skills-from-content"&gt;LinkedIn extracts skill information from texts&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;They extract skills from various content across the platform and map these skills to their Skills Graph to ensure accurate job and learning matches.&lt;/p&gt;

&lt;h2&gt;
  
  
  🗝 &lt;a href="https://security.googleblog.com/2024/04/accelerating-incident-response-using.html"&gt;Google speeds up security and privacy incidents workflows&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;They use LLMs to summarize incidents for different audiences, including executives, leads, and partner teams. It saves responders’ time and improves the quality of incident summaries. &lt;/p&gt;

&lt;h2&gt;
  
  
  🏪 &lt;a href="https://blog.picnic.nl/enhancing-search-retrieval-with-large-language-models-llms-7c3748b26d72"&gt;Picnic improves search relevance for product listings&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;They leverage LLMs to enhance product and recipe search retrieval for users from three countries with their own unique language and culinary preferences. &lt;/p&gt;

&lt;h2&gt;
  
  
  🙅 &lt;a href="https://engineeringblog.yelp.com/2024/03/ai-pipeline-inappropriate-language-detection.html"&gt;Yelp detects inappropriate language in reviews&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;The company enhanced its content moderation system with LLMs to help identify egregious instances of threats, harassment, lewdness, personal attacks, or hate speech.&lt;/p&gt;

&lt;h2&gt;
  
  
  🚗 &lt;a href="https://www.uber.com/en-GB/blog/generative-ai-for-high-quality-mobile-testing/?uclick_id=13598f30-43e0-466c-a42c-347f4bab3bbf"&gt;Uber tests mobile applications&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;They created DragonCrawl, a system that uses LLMs to execute mobile tests with the intuition of a human. It saves thousands of developer hours and reduces test maintenance costs.&lt;/p&gt;

&lt;h2&gt;
  
  
  #️⃣ &lt;a href="https://engineering.grab.com/llm-powered-data-classification"&gt;Grab automatically tags sensitive data&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;They use LLM to classify data entities, identify sensitive data, and assign the most appropriate tag to each entity. &lt;/p&gt;

&lt;h2&gt;
  
  
  🛒 &lt;a href="https://tech.instacart.com/scaling-productivity-with-ava-instacarts-internal-ai-assistant-ed7f02558d84"&gt;Instacart builds an internal AI assistant&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;Teams use an internal AI assistant called Ava to write, review and debug code, improve communications, and build AI-enabled internal tools on top of the company’s APIs.&lt;/p&gt;

&lt;h2&gt;
  
  
  🛍 &lt;a href="https://medium.com/whatnot-engineering/how-whatnot-utilizes-generative-ai-to-enhance-trust-and-safety-c7968eb6315e"&gt;Whatnot detects marketplace spam&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;They use LLMs to enhance trust and safety areas like multimodal content moderation, fulfillment, bidding irregularities, and general fraud protection. &lt;/p&gt;

&lt;h2&gt;
  
  
  💌 &lt;a href="https://engblog.nextdoor.com/let-ai-entertain-you-increasing-user-engagement-with-generative-ai-and-rejection-sampling-50a402264f56"&gt;Nextdoor generates engaging email subject lines&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;The company aims to generate informative and engaging subject lines that will lead to more email opens, clicks, and eventually more sessions on the platform.&lt;/p&gt;

&lt;h2&gt;
  
  
  🍿 &lt;a href="https://medium.com/vimeo-engineering-blog/from-idea-to-reality-elevating-our-customer-support-through-generative-ai-101a2c5ea680"&gt;Vimeo builds customer support AI assistant&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;They prototyped a help desk chatbot where customers input their questions and receive immediate, accurate, and personalized responses.&lt;/p&gt;

&lt;h2&gt;
  
  
  🤖 &lt;a href="https://www.godaddy.com/resources/news/llm-from-the-trenches-10-lessons-learned-operationalizing-models-at-godaddy"&gt;GoDaddy classifies support inquiries&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;GoDaddy leverages LLMs to enhance customer experience in their messaging channels by classifying support inquiries. They share lessons learned operationalizing these models.&lt;/p&gt;

&lt;h2&gt;
  
  
  🗞 &lt;a href="https://tech.olx.com/extracting-job-roles-in-job-ads-a-journey-with-generative-ai-e8b8cf399659"&gt;OLX extracts information from job listings&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;They use Prosus AI Assistant, their generative AI (GenAI) model, to extract job roles in job ads and ensure a closer alignment between job seekers’ desired jobs and the relevant listings.&lt;/p&gt;

&lt;h2&gt;
  
  
  🔢 &lt;a href="https://www.honeycomb.io/blog/we-shipped-ai-product"&gt;Honeycomb helps users write data queries&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;The company built Query Assistant to accelerate users’ learning curve associated with queries. Users can describe or ask things in plain English like “slow endpoints by status code” and Query Assistant will generate a relevant Honeycomb query to iterate on.&lt;/p&gt;

&lt;h2&gt;
  
  
  📦 &lt;a href="https://doordash.engineering/2024/04/23/building-doordashs-product-knowledge-graph-with-large-language-models/"&gt;DoorDash extracts product attributes from unstructured SKU data&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;They use LLMs to extract and tag product attributes from raw merchant data. It allows to easily match customer queries with relevant items on DoorDash and helps delivery drivers to find the correct product in the store. &lt;/p&gt;

&lt;h2&gt;
  
  
  ⚠️ &lt;a href="https://incident.io/blog/lessons-learned-from-building-our-first-ai-product"&gt;Incident.io generates summaries of software incidents&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;Incident.io helps to collaborate on software incidents by suggesting and updating the incident summary. This suggestion considers the latest update, the conversation in the Slack channel, and the previous summary. Half of all summary updates in Incident.io are now written by AI.&lt;/p&gt;

&lt;h2&gt;
  
  
  🪡 &lt;a href="https://multithreaded.stitchfix.com/blog/2023/03/06/expert-in-the-loop-generative-ai-at-stitch-fix/"&gt;StitchFix generates ad headlines and product descriptions&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;The company combines algo-generated text with a human expert-in-the-loop approach to streamline crafting engaging advertisement headlines and producing high-fidelity product descriptions.&lt;/p&gt;

&lt;h2&gt;
  
  
  💳 &lt;a href="https://digits.com/developer/posts/assisting-accountants-with-generative-machine-learning/"&gt;Digits suggests questions about banking transactions&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;They use generative models to assist their customers – accountants – by suggesting questions about a transaction to a client. The accountants can then send the question to their client as is, or edit it without typing everything from scratch.&lt;/p&gt;

&lt;h2&gt;
  
  
  🧑‍🏫 &lt;a href="https://blog.duolingo.com/large-language-model-duolingo-lessons/"&gt;Duolingo generates content for lessons&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;The company leverages LLMs to help their learning designers come up with relevant exercises for lessons. Human experts plan out the theme, grammar, vocabulary, and exercise types for a given lesson and the model outputs relevant exercises.&lt;/p&gt;

&lt;h2&gt;
  
  
  🏠 &lt;a href="https://www.zillow.com/tech/using-ai-to-understand-the-complexities-and-pitfalls-of-real-estate-data/"&gt;Zillow detects discriminatory content in real-estate listings&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;The company uses LLMs to understand whether real-estate listings contain proxy for race and other remnants of historical inequalities in the real estate domain.&lt;/p&gt;

&lt;h2&gt;
  
  
  🍲 &lt;a href="https://bytes.swiggy.com/improving-search-relevance-in-hyperlocal-food-delivery-using-small-language-models-ecda2acc24e6"&gt;Swiggy improves search relevance in hyperlocal food delivery&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;They use LLMs to match search queries in a variety of languages with millions of dish names with regional variety. &lt;/p&gt;

&lt;h2&gt;
  
  
  Want more examples of LLM systems in production?
&lt;/h2&gt;

&lt;p&gt;Check out our &lt;a href="https://www.evidentlyai.com/ml-system-design"&gt;database&lt;/a&gt; of 450 use cases from 100+ companies that share their learnings from implementing ML and LLM systems. Bookmark the list and enjoy the reading!&lt;/p&gt;

</description>
      <category>llm</category>
      <category>machinelearning</category>
      <category>ai</category>
      <category>genai</category>
    </item>
  </channel>
</rss>
