<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Priyanshu S</title>
    <description>The latest articles on DEV Community by Priyanshu S (@priyanshu_evaliphy).</description>
    <link>https://dev.to/priyanshu_evaliphy</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3871411%2F390040e7-5215-4db5-993c-9c41748551c2.png</url>
      <title>DEV Community: Priyanshu S</title>
      <link>https://dev.to/priyanshu_evaliphy</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/priyanshu_evaliphy"/>
    <language>en</language>
    <item>
      <title>How to Handle Proprietary Jargon in LLM-as-a-Judge Evaluations</title>
      <dc:creator>Priyanshu S</dc:creator>
      <pubDate>Sun, 26 Apr 2026 11:10:01 +0000</pubDate>
      <link>https://dev.to/evaliphy/how-to-handle-proprietary-jargon-in-llm-as-a-judge-evaluations-55d3</link>
      <guid>https://dev.to/evaliphy/how-to-handle-proprietary-jargon-in-llm-as-a-judge-evaluations-55d3</guid>
      <description>&lt;p&gt;Imagine explaining a complex medical procedure or a niche legal clause to a bright high school student. They’re smart, sure, but they lack the years of context that make your industry unique. &lt;/p&gt;

&lt;p&gt;This is exactly the challenge we face when using &lt;strong&gt;Frontier Models&lt;/strong&gt; like GPT-4 as judges for &lt;strong&gt;&lt;a href="https://developers.llamaindex.ai/python/framework/understanding/rag/" rel="noopener noreferrer"&gt;RAG (Retrieval-Augmented Generation) systems&lt;/a&gt;&lt;/strong&gt;. These models are incredibly capable, but they often default to "layman" logic. When they encounter your company's proprietary codes, specialized jargon, or industry-specific shorthand, they might get confused.&lt;/p&gt;

&lt;p&gt;Without the right strategy, your AI judge might hallucinate, falsely penalize a perfectly correct answer, or—worse—tell you everything is fine when it’s actually a mess. If your AI judge doesn't understand your domain, it isn't really a judge; it's just a guesser. (If you're new to this, check out our &lt;a href="https://evaliphy.com/blog/welcome-to-evaliphy" rel="noopener noreferrer"&gt;Introduction to Evaliphy&lt;/a&gt; to see how we're simplifying RAG testing).&lt;/p&gt;

&lt;p&gt;Here are six human-tested strategies to turn a general-purpose AI into a domain expert for your &lt;strong&gt;AI evaluation&lt;/strong&gt; workflow.&lt;/p&gt;




&lt;h2&gt;
  
  
  1. Use Reference-Based Evaluation (The "Cheat Sheet")
&lt;/h2&gt;

&lt;p&gt;The simplest way to stop an AI from guessing is to give it the answer key. Instead of asking the judge, "Is this answer correct?" (which forces it to rely on its own potentially outdated knowledge), you provide a &lt;strong&gt;Reference Answer&lt;/strong&gt; (Ground Truth).&lt;/p&gt;

&lt;p&gt;Think of it like a semantic matching game. You ask the judge: "Does the model's answer mean the same thing as this verified Reference Answer?" &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Example of Reference-Based Evaluation:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;The Jargon:&lt;/strong&gt; "The system requires a &lt;em&gt;Cold-Start Reboot&lt;/em&gt; via the &lt;em&gt;Alpha-9&lt;/em&gt; protocol."&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;The AI's Answer:&lt;/strong&gt; "You need to restart the computer using the standard menu."&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;The Reference Answer:&lt;/strong&gt; "Perform a hard power cycle and initiate the Alpha-9 sequence."&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;The Result:&lt;/strong&gt; By comparing the two, the judge can see the AI's answer is too generic and misses the critical "Alpha-9" step, even if it doesn't know what Alpha-9 is.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  2. Implement Few-Shot Prompting (Show, Don't Just Tell)
&lt;/h2&gt;

&lt;p&gt;AI models are great mimics. If you want the judge to understand how to handle your specific jargon, show it a few examples of what a "good" and "bad" answer looks like in your world. This is known as &lt;strong&gt;In-Context Learning&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Example of Few-Shot Prompting:&lt;/strong&gt;&lt;br&gt;
Provide the judge with three pairs like this in your prompt:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Query:&lt;/strong&gt; "How do I handle a &lt;em&gt;Tier-3&lt;/em&gt; escalation?"&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Context:&lt;/strong&gt; "Tier-3 escalations must be logged in the &lt;em&gt;Red-Book&lt;/em&gt;."&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Good Answer:&lt;/strong&gt; "Log it in the Red-Book." (Score: 5)&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Bad Answer:&lt;/strong&gt; "Tell your manager." (Score: 1 - Reason: Failed to mention the Red-Book requirement).&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;The Result:&lt;/strong&gt; After seeing these examples, the judge learns that "Red-Book" is a non-negotiable term in your workflow.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  3. Define Detailed Rubrics and Criteria
&lt;/h2&gt;

&lt;p&gt;Vague instructions lead to vague results. If you tell a judge to check for "helpfulness," it will use its own definition of helpful. To ensure &lt;strong&gt;accurate AI evaluation&lt;/strong&gt;, you need to define your terms explicitly. For more advanced cases, you can even &lt;a href="https://evaliphy.com/blog/tuning-llm-judge-custom-prompts" rel="noopener noreferrer"&gt;tune your LLM judge with custom prompts&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Example of an Evaluation Rubric:&lt;/strong&gt;&lt;br&gt;
Instead of "Is the answer accurate?", use a rubric like this:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Score 5 (Expert):&lt;/strong&gt; Uses the term &lt;em&gt;'Fiduciary Duty'&lt;/em&gt; correctly and mentions the &lt;em&gt;'2024 Compliance Update'&lt;/em&gt;.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Score 3 (Layman):&lt;/strong&gt; Explains the concept of duty but misses the specific 2024 update.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Score 1 (Incorrect):&lt;/strong&gt; Suggests the user has no legal obligation.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;The Result:&lt;/strong&gt; The judge now has a clear checklist to follow, reducing the chance of it being "too nice" to a mediocre answer.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  4. Fine-Tune a Specialized Judge Model
&lt;/h2&gt;

&lt;p&gt;Sometimes, a general-purpose model is just too "general." If your industry is drowning in thousands of unique codes, it might be time to build your own specialist.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Example of Fine-Tuning for Jargon:&lt;/strong&gt;&lt;br&gt;
A medical tech company might take a base model like Llama-3 and train it on 5,000 examples of their internal hardware error codes (e.g., "Error E-112: Oxygen Sensor Desaturation"). &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;The Result:&lt;/strong&gt; The fine-tuned model becomes a &lt;strong&gt;Custom Judge&lt;/strong&gt; that recognizes "E-112" instantly, whereas a generic model might think it's just a random typo.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  5. Calibrate with Subject Matter Experts (SME)
&lt;/h2&gt;

&lt;p&gt;Even the best AI needs a reality check. Research shows that AI judges often agree with "regular people" about 80% of the time, but their agreement with actual experts (like doctors or lawyers) can drop as low as 60%. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Example of SME Calibration:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;The Scenario:&lt;/strong&gt; An AI judge gives a "Pass" to a legal summary.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;The SME Review:&lt;/strong&gt; A senior lawyer looks at it and says, "Wait, the AI missed that this clause only applies in &lt;em&gt;Delaware&lt;/em&gt; law, not &lt;em&gt;New York&lt;/em&gt;."&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;The Fix:&lt;/strong&gt; You take that specific "Delaware vs. New York" example and add it to your Few-Shot examples (Strategy 2).&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;The Result:&lt;/strong&gt; The AI judge gets smarter with every human correction.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  6. Focus on Grounding (The "Paper Trail")
&lt;/h2&gt;

&lt;p&gt;If you’re worried about the AI hallucinating facts about your jargon, change the question. Instead of asking "Is this factually true?", ask "Is this answer supported &lt;em&gt;only&lt;/em&gt; by the provided text?" This is often called &lt;strong&gt;Faithfulness&lt;/strong&gt; or &lt;strong&gt;Groundedness&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Example of Grounding Check:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;The Source Text:&lt;/strong&gt; "The &lt;em&gt;XJ-900&lt;/em&gt; unit must be oiled every 4 hours."&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;The AI's Answer:&lt;/strong&gt; "The XJ-900 is a high-performance engine that needs daily maintenance."&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;The Judge's Task:&lt;/strong&gt; "Does the source text say it's a high-performance engine? Does it say daily maintenance?"&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;The Result:&lt;/strong&gt; The judge fails the answer because the source text &lt;em&gt;only&lt;/em&gt; mentioned the 4-hour oiling schedule. It doesn't matter if the AI's "daily" guess is technically true in the real world; it failed the "Paper Trail" test.&lt;/li&gt;
&lt;/ul&gt;




&lt;h3&gt;
  
  
  Conclusion: Building Reliable RAG Evaluations
&lt;/h3&gt;

&lt;p&gt;Jargon shouldn't be a barrier to building great AI. By using these strategies, you move away from "black box" testing and toward a system where your evaluations are as specialized as your business.&lt;/p&gt;

&lt;p&gt;At &lt;strong&gt;&lt;a href="https://evaliphy.com" rel="noopener noreferrer"&gt;Evaliphy&lt;/a&gt;&lt;/strong&gt;, we believe that the goal isn't just to have an AI that talks; it's to have an AI that truly understands what you're saying. By implementing reference-based checks and clear rubrics, you can ensure your RAG system remains accurate, even in the most specialized domains.&lt;/p&gt;




</description>
      <category>aieval</category>
      <category>rag</category>
      <category>llm</category>
      <category>testing</category>
    </item>
    <item>
      <title>Simplifying the AI Testing through Evaliphy</title>
      <dc:creator>Priyanshu S</dc:creator>
      <pubDate>Fri, 10 Apr 2026 10:01:39 +0000</pubDate>
      <link>https://dev.to/priyanshu_evaliphy/simplifying-the-ai-testing-through-evaliphy-4h0k</link>
      <guid>https://dev.to/priyanshu_evaliphy/simplifying-the-ai-testing-through-evaliphy-4h0k</guid>
      <description>&lt;p&gt;I have been a QA consultant for more than a decade. I test software for a living. APIs, UIs, integrations, if they break, I have probably written a test for it at some point.&lt;/p&gt;

&lt;p&gt;A few months ago, I started building something called &lt;a href="https://evaliphy.com" rel="noopener noreferrer"&gt;Evaliphy&lt;/a&gt; - A SDK for AI testing.&lt;/p&gt;

&lt;p&gt;Not because I set out to build a tool. But because I spotted a gap.&lt;/p&gt;

&lt;h2&gt;
  
  
  The context
&lt;/h2&gt;

&lt;p&gt;Last few months, I have been exploring and diving into AI space to update myself. At the same time, the teams I was working with started with AI features. RAG being the first feature we delivered a year back, and we casually shipped it through Vibe-Testing.&lt;/p&gt;

&lt;p&gt;A few months later, the same RAG feature which we shipped as beta, was ready to scale and go live for all the customers. &lt;/p&gt;

&lt;p&gt;The causal vibe-testing which we did a year back was not any longer an effective way to test, and honestly we did not know how to test non-deterministic systems.&lt;/p&gt;

&lt;p&gt;The only option which we had at that point of time, was to learn AI evaluation. That's what we did. &lt;/p&gt;

&lt;p&gt;We explored tools like DeepEval, Ragas, Promptfoo, etc. &lt;/p&gt;

&lt;p&gt;All of these are established and super powerful tools, we decided to use Promptfoo and built an evaluation suite in that.&lt;/p&gt;




&lt;h2&gt;
  
  
  The exploration phase
&lt;/h2&gt;

&lt;p&gt;These tools are genuinely impressive. The people who built them clearly understand AI evaluation deeply. The metrics they expose — faithfulness scores, retrieval precision, context recall — are grounded in real ML research.&lt;/p&gt;

&lt;p&gt;But at some point of time, I stopped and asked myself an honest question.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Who is this actually built for?&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Because it was not built for me. Not for the version of me that spends most of my working life writing Playwright tests, treating systems as black boxes, validating what users actually experience. &lt;/p&gt;

&lt;p&gt;I kept hitting the same wall. To use these tools properly I needed access to retrieval pipeline internals. I needed to understand embedding similarity. I needed to construct test cases based on internals of the system, providing retrieved context that I, as a black box QA engineer, simply did not want to. &lt;br&gt;
[Note: Its not skill gap, rather a testing risk because we are testing system from LLM perspective, but ignoring end-to-end tests which eventually matter more for end users.]&lt;/p&gt;


&lt;h2&gt;
  
  
  The gap nobody was talking about
&lt;/h2&gt;

&lt;p&gt;Here is what I noticed. Every article, every tutorial, every conference talk about AI evaluation was aimed at ML engineers and developers who built the AI system. The assumption was always that the person doing the evaluation had built the thing, understood the internals, and could instrument the pipeline.&lt;/p&gt;

&lt;p&gt;But that is not how QA works.&lt;/p&gt;

&lt;p&gt;In every team I have ever worked with, QA engineers own end-to-end black box testing. They test the behaviour the user sees. They do not need to understand how the database query optimizer works to test whether search results are correct. They do not need to understand OAuth internals to test whether login works.&lt;/p&gt;

&lt;p&gt;That separation exists for good reasons. Independent testing means independent perspective. A QA engineer who did not build the system will catch things the developer never will, precisely because they are not anchored to how it is supposed to work internally.&lt;/p&gt;

&lt;p&gt;But for AI features, that separation had completely broken down. The only people testing AI quality were the people who built it. Using tools that required them to build it.&lt;/p&gt;

&lt;p&gt;I found that genuinely concerning.&lt;/p&gt;


&lt;h2&gt;
  
  
  What I was trying to build
&lt;/h2&gt;

&lt;p&gt;I did not want to replace DeepEval or Ragas. Those tools solve real problems for the right audience. What I wanted was something that felt like Playwright — familiar, approachable, built around the workflows QA engineers already know.&lt;/p&gt;

&lt;p&gt;The core idea was simple. A QA engineer should be able to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Point the tool at an HTTP endpoint&lt;/li&gt;
&lt;li&gt;Write an assertion in plain English&lt;/li&gt;
&lt;li&gt;Get a pass or fail result&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;No Python environment. No pipeline access. No ML concepts to learn before writing the first test. Just an endpoint and an assertion.&lt;/p&gt;

&lt;p&gt;I wanted the entry point to be &lt;code&gt;npm install&lt;/code&gt; and the first test to take fifteen minutes to write, not three days.&lt;/p&gt;

&lt;p&gt;And, thats how I built "&lt;a href="https://evaliphy.com" rel="noopener noreferrer"&gt;Evaliphy&lt;/a&gt;"&lt;/p&gt;


&lt;h2&gt;
  
  
  What Evaliphy looks like today
&lt;/h2&gt;

&lt;p&gt;After weeks of building, iterating, breaking things, and fixing them, Evaliphy is in beta.&lt;/p&gt;

&lt;p&gt;Here is what a basic evaluation looks like:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight typescript"&gt;&lt;code&gt;&lt;span class="k"&gt;import&lt;/span&gt; &lt;span class="p"&gt;{&lt;/span&gt; &lt;span class="nx"&gt;evaluate&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="nx"&gt;expect&lt;/span&gt; &lt;span class="p"&gt;}&lt;/span&gt; &lt;span class="k"&gt;from&lt;/span&gt; &lt;span class="dl"&gt;'&lt;/span&gt;&lt;span class="s1"&gt;evaliphy&lt;/span&gt;&lt;span class="dl"&gt;'&lt;/span&gt;&lt;span class="p"&gt;;&lt;/span&gt;

&lt;span class="nf"&gt;evaluate&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="dl"&gt;'&lt;/span&gt;&lt;span class="s1"&gt;return policy check&lt;/span&gt;&lt;span class="dl"&gt;'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="k"&gt;async &lt;/span&gt;&lt;span class="p"&gt;({&lt;/span&gt; &lt;span class="nx"&gt;httpClient&lt;/span&gt; &lt;span class="p"&gt;})&lt;/span&gt; &lt;span class="o"&gt;=&amp;gt;&lt;/span&gt; &lt;span class="p"&gt;{&lt;/span&gt;
  &lt;span class="kd"&gt;const&lt;/span&gt; &lt;span class="nx"&gt;res&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="k"&gt;await&lt;/span&gt; &lt;span class="nx"&gt;httpClient&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nf"&gt;post&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="dl"&gt;'&lt;/span&gt;&lt;span class="s1"&gt;/api/chat&lt;/span&gt;&lt;span class="dl"&gt;'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="p"&gt;{&lt;/span&gt;
    &lt;span class="na"&gt;message&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="dl"&gt;'&lt;/span&gt;&lt;span class="s1"&gt;What is your return policy?&lt;/span&gt;&lt;span class="dl"&gt;'&lt;/span&gt;
  &lt;span class="p"&gt;});&lt;/span&gt;

  &lt;span class="kd"&gt;const&lt;/span&gt; &lt;span class="p"&gt;{&lt;/span&gt; &lt;span class="nx"&gt;answer&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="nx"&gt;context&lt;/span&gt; &lt;span class="p"&gt;}&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="k"&gt;await&lt;/span&gt; &lt;span class="nx"&gt;res&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nf"&gt;json&lt;/span&gt;&lt;span class="p"&gt;();&lt;/span&gt;

  &lt;span class="k"&gt;await&lt;/span&gt; &lt;span class="nf"&gt;expect&lt;/span&gt;&lt;span class="p"&gt;({&lt;/span&gt;
    &lt;span class="na"&gt;query&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="dl"&gt;'&lt;/span&gt;&lt;span class="s1"&gt;What is your return policy?&lt;/span&gt;&lt;span class="dl"&gt;'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
    &lt;span class="na"&gt;response&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="nx"&gt;answer&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
    &lt;span class="nx"&gt;context&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
  &lt;span class="p"&gt;}).&lt;/span&gt;&lt;span class="nf"&gt;toBeFaithful&lt;/span&gt;&lt;span class="p"&gt;();&lt;/span&gt;

  &lt;span class="k"&gt;await&lt;/span&gt; &lt;span class="nf"&gt;expect&lt;/span&gt;&lt;span class="p"&gt;({&lt;/span&gt;
    &lt;span class="na"&gt;query&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="dl"&gt;'&lt;/span&gt;&lt;span class="s1"&gt;What is your return policy?&lt;/span&gt;&lt;span class="dl"&gt;'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
    &lt;span class="na"&gt;response&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt; &lt;span class="nx"&gt;answer&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
    &lt;span class="nx"&gt;context&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
  &lt;span class="p"&gt;}).&lt;/span&gt;&lt;span class="nf"&gt;toBeRelevant&lt;/span&gt;&lt;span class="p"&gt;();&lt;/span&gt;
&lt;span class="p"&gt;});&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;That is it. You point it at your API. You write assertions. You run it.&lt;/p&gt;

&lt;p&gt;No Python. No pipeline access. No ML background required. If you have written a Playwright test before, this will feel familiar within minutes.&lt;/p&gt;

&lt;p&gt;The built-in judge handles the LLM evaluation underneath. The prompts are shipped with the SDK. You configure your API key and base URL, and everything else is taken care of.&lt;/p&gt;




&lt;h2&gt;
  
  
  The honest limitations
&lt;/h2&gt;

&lt;p&gt;I want to be clear about something because I think honesty matters more than marketing.&lt;/p&gt;

&lt;p&gt;Evaliphy is not a replacement for deep AI evaluation. If you are an ML engineer who wants to measure retrieval precision, chunk quality, and embedding similarity — DeepEval and Ragas are still the right tools for that job.&lt;/p&gt;

&lt;p&gt;What Evaliphy gives you is a starting point. Meaningful coverage today, from the QA engineer's perspective, without the prerequisite of ML expertise.&lt;/p&gt;

&lt;p&gt;Does that mean you never need to understand the internals? No. Deep knowledge always improves testing. It does for every system. But should that knowledge be the entry ticket before a QA engineer can write their first test? I do not think so.&lt;/p&gt;

&lt;p&gt;We do not ask QA engineers to understand database internals before testing search functionality. We should not ask them to understand retrieval pipelines before testing AI responses.&lt;/p&gt;




&lt;h2&gt;
  
  
  Where it goes from here
&lt;/h2&gt;

&lt;p&gt;Evaliphy is open source and completely free. It is in beta which means the API will change, there will be rough edges, and your feedback will directly shape what it becomes.&lt;/p&gt;

&lt;p&gt;The vision is bigger than RAG. RAG is where the pain is most acute right now but the same problem exists across AI systems — there is no standard black box E2E testing tool for any of them. That is what I want to build toward.&lt;/p&gt;

&lt;p&gt;If you are a QA engineer who has been handed an AI feature and is not sure where to start — Evaliphy is for you.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;npm &lt;span class="nb"&gt;install &lt;/span&gt;evaliphy
evaliphy init my-first-eval
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;And if you try it and something does not work, open an issue. If an assertion you need does not exist, tell me. If the whole approach feels wrong, I want to know that too.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Evaliphy is open source and in beta. You can find it at &lt;a href="https://evaliphy.com" rel="noopener noreferrer"&gt;evaliphy.com&lt;/a&gt; or install it directly with &lt;code&gt;npm install evaliphy&lt;/code&gt;. Honest feedback is more valuable than stars right now.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>testing</category>
      <category>rag</category>
      <category>llm</category>
    </item>
  </channel>
</rss>
