ZenML

AI-Powered Call Center Agents for Healthcare Operations

HeyRevia 2023
View original source

HeyRevia has developed an AI call center solution specifically for healthcare operations, where over 30% of operations run on phone calls. Their system uses AI agents to handle complex healthcare-related calls, including insurance verifications, prior authorizations, and claims processing. The solution incorporates real-time audio processing, context understanding, and sophisticated planning capabilities to achieve performance that reportedly exceeds human operators while maintaining compliance with healthcare regulations.

Industry

Healthcare

Technologies

Overview

HeyRevia is building AI-powered call center agents specifically designed for the healthcare industry. The presentation was given by Sean, the company’s CEO, who brings a decade of AI experience including work on Google Assistant (notably the 2018 AI calling demonstration for restaurants and salons) and autonomous vehicle development at Waymo. This background in both conversational AI and autonomous systems heavily influences their approach to healthcare call center automation.

The core problem HeyRevia addresses is that more than 30% of healthcare operations still run through phone calls. These calls span a wide range of activities from simple appointment scheduling to complex negotiations with insurance companies regarding credential verification, prior authorizations, referrals, claims denials, and benefits inquiries. The current industry solution involves Business Process Outsourcing (BPO) providers where human agents in call centers often end up calling each other, sometimes sitting in adjacent rooms but still required to communicate via phone. This represents a significant inefficiency that HeyRevia aims to solve with AI agents.

Voice Agent Landscape and Technical Challenges

Sean provides valuable context on the current state of voice agent technology. Over the past two years, speech-to-text (STT) and text-to-speech (TTS) capabilities have improved dramatically, and large language models have evolved from text-only inputs to handling audio directly, as demonstrated by OpenAI’s real-time API. However, significant production challenges remain.

The typical voice agent architecture follows a pipeline approach used by platforms like VAPI, Retell, and Bland. This pipeline flows from telephony systems (like Twilio or 10x) through streaming input handling via WebSocket and WebRTC, to ASR solutions (Assembly AI, Deepgram, Whisper) for speech-to-text conversion, then to the LLM for understanding and response generation, and finally back through TTS to produce audio output.

The key limitations of this pipeline approach include:

HeyRevia’s Architecture: Perception-Prediction-Planning-Control

HeyRevia’s architecture draws significant inspiration from autonomous vehicle systems, which is unsurprising given Sean’s background at Waymo. Rather than treating voice interaction as a simple pipeline, they model it as an autonomous agent operating in an environment with multiple states and required behaviors.

Perception Layer

The perception layer continuously processes incoming audio to understand the current state of the call. It can distinguish between:

This real-time perception allows the system to adapt its behavior appropriately rather than processing all audio uniformly.

Prediction Layer

The prediction layer anticipates what should happen next. A critical optimization mentioned is hold handling: when the system detects that a call is on hold (through the perception layer), it pauses all processing and LLM inference for that call. The agent “sits silently” and waits for a human to join. This saves significant token costs during what could be 30-minute hold times while ensuring the system is ready to respond immediately when a human representative joins.

Planning Layer

The planning layer addresses what Sean identifies as the primary difference between voice agents and humans: the ability to think ahead. With simple prompt-based approaches, there’s no mechanism to provide the AI with sequenced, contextual information about what needs to happen at specific points in the call. The planning layer enables the agent to:

Control Layer

The control layer provides guardrails to prevent the agent from going off-track. This is explicitly designed to prevent hallucination and scope creep. For example, when working with pharmaceutical companies, the control layer ensures the AI stays focused on medical information and doesn’t drift into irrelevant topics like discussing meals or lunch.

Operational Features

Human-in-the-Loop Capability

A distinctive feature of HeyRevia is the ability for human supervisors to take over calls in real-time. When monitoring multiple simultaneous calls (the presentation shows 10-15 concurrent calls), a supervisor can:

This provides both quality assurance and a recovery mechanism for edge cases the AI cannot handle.

Call Center API vs. UI

HeyRevia offers two integration patterns:

Work API (Call Center API): This treats the AI as a task executor. Users submit call work items, and the AI handles them autonomously. Importantly, the system has self-correction capabilities - if a call fails due to missing or incorrect information (like an invalid provider ID or NPI number), the AI can identify the issue and request the correct information before retrying. This represents the AI “learning from its mistakes.”

Call Center UI: Provides a visual interface for monitoring and intervening in calls, enabling the human-in-the-loop functionality described above.

Evaluation and Benchmarking

HeyRevia’s evaluation philosophy is “if you’re trying to ask AI agent to do similar human work, you have to evaluate it like a human.” They benchmark AI performance against human agents on the same scenarios by analyzing transcripts and comparing outcomes. According to their data, the AI outperforms humans in many scenarios.

A concrete example provided: for insurance claims where the initial claim was denied, human agents typically require two to three phone calls to identify the actual denial reason, while their AI can achieve this in one to two calls by more effectively negotiating with and pushing back on human representatives.

However, Sean acknowledges that LLMs “do make simple and stupid mistakes” - the challenge is catching and handling these during live calls, which is addressed through the control layer and human intervention capabilities.

Healthcare Compliance and Production Considerations

Operating in healthcare requires extensive compliance measures:

EHR Integration

Currently, HeyRevia does not directly integrate with Electronic Health Record (EHR) systems. They operate as a layer on top, functioning as an AI call center that works on behalf of customers. Direct EHR integration may come as the company matures and demonstrates “proof of work.”

Real-World Use Cases

The system handles common healthcare phone-based workflows including:

Each call type involves navigating IVR systems, providing repeated identifying information (NPI numbers, member IDs, etc.), waiting on hold, and then negotiating with human representatives - all of which the AI can handle while humans previously had to dedicate significant time to these tedious processes.

Production Insights and Lessons

Several practical insights emerge from this case study:

The case study represents an interesting application of autonomous agent principles to a highly regulated, high-stakes domain where the consequences of errors are severe but the potential for efficiency gains is substantial.

More Like This

Agentic AI Copilot for Insurance Underwriting with Multi-Tool Integration

Snorkel 2025

Snorkel developed a specialized benchmark dataset for evaluating AI agents in insurance underwriting, leveraging their expert network of Chartered Property and Casualty Underwriters (CPCUs). The benchmark simulates an AI copilot that assists junior underwriters by reasoning over proprietary knowledge, using multiple tools including databases and underwriting guidelines, and engaging in multi-turn conversations. The evaluation revealed significant performance variations across frontier models (single digits to ~80% accuracy), with notable error modes including tool use failures (36% of conversations) and hallucinations from pretrained domain knowledge, particularly from OpenAI models which hallucinated non-existent insurance products 15-45% of the time.

healthcare fraud_detection customer_support +90

Building and Scaling Conversational Voice AI Agents for Enterprise Go-to-Market

Thoughtly / Gladia 2025

Thoughtly, a voice AI platform founded in late 2023, provides conversational AI agents for enterprise sales and customer support operations. The company orchestrates speech-to-text, large language models, and text-to-speech systems to handle millions of voice calls with sub-second latency requirements. By optimizing every layer of their stack—from telephony providers to LLM inference—and implementing sophisticated caching, conditional navigation, and evaluation frameworks, Thoughtly delivers 3x conversion rates over traditional methods and 15x ROI for customers. The platform serves enterprises with HIPAA and SOC 2 compliance while handling both inbound customer support and outbound lead activation at massive scale across multiple languages and regions.

customer_support healthcare regulatory_compliance +33

Panel Discussion on LLMOps Challenges: Model Selection, Ethics, and Production Deployment

Google, Databricks, 2023

A panel discussion featuring leaders from various AI companies discussing the challenges and solutions in deploying LLMs in production. Key topics included model selection criteria, cost optimization, ethical considerations, and architectural decisions. The discussion highlighted practical experiences from companies like Interact.ai's healthcare deployment, Inflection AI's emotionally intelligent models, and insights from Google and Databricks on responsible AI deployment and tooling.

healthcare customer_support high_stakes_application +27