DEV Community

Cover image for AI Agents vs. Real Customers: What Could Possibly Go Wrong?
MCnad
MCnad

Posted on

5 2 2 3 4

AI Agents vs. Real Customers: What Could Possibly Go Wrong?

Today, a 70% accuracy rate is often considered a success for large language models (LLMs) in human interactions. But, for well-established brands, this threshold introduces serious reputational and legal risks. So, how close are we to truly reliable, AI-driven customer interactions?

A conversation between LLM and a human user showing the user asks

AI alignment isn’t just about tuning a model; it’s about ensuring that an AI chat agent, for example, consistently conforms to a company’s needs (and rules) and, just as importantly, serves its customers effectively. Unlike human support agents, who can quickly adapt and clarify misunderstandings, LLMs can spiral into frustrating loops, expose sensitive information, or completely misinterpret user intent if they aren’t properly aligned.

Is Technology Ready?

The LLM revolution accelerated with the introduction of Transformer models in the monumental paper, Attention Is All You Need (2017), paving the way to advanced AI applications like ChatGPT and Claude.

But is this technology truly ready to autonomously handle customer interactions? The short answer: Yes. But it’s not just about AI capabilities; it’s about how LLMs are utilized as part of a larger solution architecture. Even when LLMs are given explicit context and instructions tailored to one’s needs, this data must be meticulously structured in a way that’s adapted to a model’s inherent attributes and tendencies. Without a robust framework, misalignment and hallucinations are inevitable.

Humans navigate conversations seamlessly because we intuitively understand context, intent, and social norms. LLMs, however, struggle with consistency even when given extensive instructions and guidelines (sometimes even more).

More Reasons Why Current AI Methodologies Fall Short When Facing Customers

Most LLM applications today prioritize efficiency and response speed over implementing mechanisms for accuracy and consistency. This creates several challenges in complex use cases, such as customer service:

  1. LLMs struggle with complex reasoning. Without additional alignment and real-time evaluation mechanisms, they easily lose focus when given more than a few explicit instructions.
  2. Conflicting instructions lead to inconsistencies. LLMs aren’t inherently good or consistent at resolving priority conflicts in instructions within prompts.

One common approach to handling these challenges is to break an LLM application's architecture into structured flowcharts. Each stage in a customer service interaction is then guided by a specialized prompt. However, this approach ironically degrades the customer experience, making LLMs feel no different from older flow-based chatbots.
The reason lies in the inherent limitations of flowchart-based structures:

  1. Intent detection is unreliable. Customers often have multiple, evolving intents that require dynamic handling rather than rigid, singular classification.
  2. Guiding the conversation is impractical. Many businesses would want AI agents to proactively shape conversations and guide users, such as countering with questions rather than immediately satisfying requests, but intent-based models aren’t well-suited for this level of experience.
  3. Context-switching is unnatural. Flow-based execution struggles to maintain coherent conversations when users switch between different topics or tasks. This can lead to interactions that feel disjointed and out of touch and, consequently, to poor customer experience.

This is why even AI customer service solutions with a 70% accuracy rate are often considered “successful” today by AI vendors. But in real-world deployments, that standard is far too low.

We Can’t Settle for 70% Accuracy

For enterprises managing one million conversations daily, 70% accuracy means that 300,000 conversations are not handled reliably! Inside this misaligned 30%, some of the mistakes can be vital, including violating company policy, providing false facts like communicating a wrong account balance for a bank client or even violating regulations.

As previously described, LLM is not a human brain, and we need to understand how to better feed it, utilize it, and optimize it.

Treating LLMs as One Component of a Larger System

At Parlant, an open-source guidance framework for customer-facing LLM agents, we approach the problem differently. Instead of relying solely on LLMs to process and generate responses, we integrate them into a broader AI system with multiple moving parts. This methodology enables:

  • Dynamic instruction filtering. In real-world deployments, AI agents must handle dozens to hundreds of instructions. Our system sorts and prioritizes only the most relevant ones for each conversation, keeping the model focused on what it really needs to do at any given point.

  • Self-critique and prioritization mechanisms. Standard LLMs prioritize instructions and information placed later in a prompt, often ignoring earlier context. Our approach introduces Attentive Reasoning Queries (stay tuned for the upcoming research paper publication), which dynamically refocus the model’s attention, ensuring all critical guidelines are applied consistently.

  • Feedback loop for continuous improvement. Instead of only evaluating model outputs in retrospect, our system analyzes in real-time how well the model adhered to each and every instruction, not only giving operators crucial feedback and insights on the model’s interpretations but also improving the model’s ability to bounce back from what could have been misguided responses to the customer.

By treating LLMs as part of a larger engine rather than the engine itself, Parlant is able to significantly improve reliability and accuracy in customer-facing scenarios.

Maximizing Alignment to Minimize Risk

If companies want truly autonomous, reliable AI-driven customer interactions, they can’t settle for a 70% success rate. At Parlant, we’re pushing the boundaries of AI alignment to set a new benchmark for accuracy and trustworthiness.
Want to learn more? Feel free to explore our open source at https://github.com/emcie-co/parlant.

Heroku

This site is built on Heroku

Join the ranks of developers at Salesforce, Airbase, DEV, and more who deploy their mission critical applications on Heroku. Sign up today and launch your first app!

Get Started

Top comments (4)

Collapse
 
dorzo profile image
DorZo

Well written, Nadav!

The big problem here, in my opinion, is that LLM applications like ChatGPT create a “wow” effect, making certain tasks seem easily achievable. However, there’s a significant gap between an impressive demo or happy-path use case and building a truly scalable, production-grade application.

At Parlant, we’re working to bridge this gap and make it a reality!

Collapse
 
karadza profile image
Juraj

...it resets every time you microwave a spoon
😂😂

Collapse
 
eyalrin profile image
Eyal Ringort

Great article! 🔥🔥

Collapse
 
agenticworkers profile image
Agentic Workers

Good post!

The Most Contextual AI Development Assistant

Pieces.app image

Our centralized storage agent works on-device, unifying various developer tools to proactively capture and enrich useful materials, streamline collaboration, and solve complex problems through a contextual understanding of your unique workflow.

👥 Ideal for solo developers, teams, and cross-company projects

Learn more