Blueprint Logo

The AI Agent Blueprint is a strategic map for launching and scaling AI in customer service.

It helps customer service, CX, and AI transformation leaders deploy fast, scale with confidence, and achieve meaningful business transformation with AI.

2

Launch it

2.1

AI fundamentals:
An introduction to AI Agents for customer service

What is an AI Agent for customer service?

A fully autonomous customer service agent that can manage all customer interactions from start to finish – without human intervention – using natural language, business context, and product knowledge. AI Agents can seek information, make decisions, and take action to comprehensively resolve even the most complex queries and requests.

Why are they different from chatbots?

AI Agents represent a major transformation from old-school chatbots. They don’t deflect, they resolve. They’re conversational, engaging, intelligent, and able to adapt to the context of every customer’s situation. They’re always on and instantly available, driving high-quality outcomes for customers at any scale. They can also take action, like issuing refunds, updating orders, and changing account settings.

Customers get what they need, fast. Support teams get time back to focus on more strategically important work. Businesses get support that scales without scaling cost. They mark a shift from automation being just a cost-saving tool to automation as a resolution engine that drives real outcomes and value. When integrated well, they elevate the entire support experience and allow teams to rethink the role – and value – of support in the business.

Benefits of using an AI Agent

The key benefits of adopting an AI Agent revolve around customer experience and operational efficiency:

  • 24/7 supportAI Agents are available around the clock, in any language, on any channel. This ensures your customers receive timely, conversational responses, regardless of where they are in the world.
  • Fast resolutions at scaleCustomers don’t need to wait hours, days, or even weeks to get an answer or resolution to their question.
  • Increased self-serviceAdding more ways for customers to get answers on their own at any time and on any channel boosts satisfaction, shortens time to resolution, and helps drive activation by making high-quality support instantly accessible.
  • Enhanced agent efficiencyBy enabling the AI Agent to handle the bulk of questions, you can do more with less, and at a lower cost. Human agents can dedicate time to the highest value interactions, optimizing the system, and even shift focus to other areas like customer success.
  • Competitive advantageAI Agents provide a differentiated support experience, so you can offer better support than your competitors and drive customer happiness, loyalty, and retention.

AI Agents create significant business value: reduced support costs, increased resolution speed, and improved customer experiences. They’re a strategic investment for modern support teams.

How do AI Agents work?

AI Agents:

  • Deeply understand customer intentUsing large language models (LLMs) and advanced natural language understanding.
  • Resolve queries and personalize interactionsUsing customer data, conversation history, and real-time context.
  • Generate contextually accurate answersUsing a technique called retrieval-augmented generation (RAG), AI Agents pull relevant content from your support knowledge base and other defined sources to craft precise, brand-aligned responses.
  • Know when to hand over to a humanWhen AI Agents aren’t able to resolve a query, they know the right time to get a human in the loop and can seamlessly hand conversations over to the right teammate.
  • Continuously improve through feedbackThey create feedback loops, surfacing content gaps and failure points so teams can close them proactively. This reflects a closed-loop system where AI is both operational and diagnostic.

It’s important to note that not all AI Agents are built the same way. For example, Fin is built on a proprietary Fin AI Engine™, which is a patented AI architecture purpose-built for the scale and complexity of customer service. It includes:

  • A custom retrieval-augmented generation (RAG) system, which understands the question, identifies, and ranks content by accuracy and relevance.
  • Multi-stage validation, which checks every answer for quality, accuracy, and policy fit.
  • Modular sub-agent architecture, which breaks each query into specialized tasks, each handled by a tailored LLM sub-agent to deliver the highest quality answers.
2.2

Getting started with AI

There isn’t a one-size-fits-all approach to AI Agent deployment. Factors like your industry, company size, and customer type will influence the route you take. But regardless, the first step is to set the AI Agent live.

Two possible approaches here are:

Test and deploy AI

This is the fast-start approach: quickly find a business problem you need to solve, set the AI Agent live, find the limits, prove value now, and build buy-in through results.AI Agents can resolve huge volumes of support queries immediately. That unlocks capacity, improves CSAT, and enables you to achieve fast time-to-value and build a business case using real outcomes.Learn how to evaluate and deploy an AI Agent in a matter of days or weeks.

Build a business case for AI

Some companies have specific criteria for deploying new software that need to be satisfied before it goes live: things like C-suite approvals and finance reviews.This blueprint will show you how to build that business case, identify quick-win opportunities, go live quickly, and measure real-world impact so you can build momentum and make the case for what comes next.
2.3

Step 1:
Build a business case
and quantify ROI


1. Identify high-value, quick-win use cases for AI

Start by aligning on goals for AI

Most teams approach AI to solve for short-term pain: fewer queues, less overtime, more breathing room when people are out.

But the opportunity is much bigger than that. Imagine the majority of your conversations were being resolved automatically: 24/7, in any language, on any channel, with high quality, and for a low cost. What would that do for your business?

You wouldn’t just be reducing pressure on the team, you’d be resetting how they work. Not just saving costs, but changing the cost structure. Not just moving faster, but offering a completely different customer experience.

That’s the bar. Set your goals with that in mind. Ask:

  • What problems are you trying to solve?
  • How do those goals support broader company objectives – whether that’s scaling without additional headcount or improving the customer experience (or both)?
  • How will you measure success? It could be resolution rate, CSAT, time saved, or other metrics that tie directly to your team’s objectives.

Defining this early helps ensure alignment and makes it easier to prove value later on.

Honeybook

The key to getting buy-in was to take a very clear and thoughtful approach, tying the AI tooling to specific team goals and challenges, and showing through initial testing how it could address them.

Elli NeeldSenior Product Education Content Creator at Honeybook
Elli Neeld

Analyze your existing conversation data to find high-value or quick-win use cases to automate

Ask:

  • What questions do we see again and again?
  • What % of support volume is repetitive or predictable?
  • Where are the biggest delays happening that AI could speed up?
  • What gets handed over to human agents that shouldn’t be?
  • What questions could be easily answered with content you already have available?

Involve your team. Their insights can help pinpoint where improvements will have the biggest impact.

Look for workflows where AI can either fully automate resolution, or significantly reduce human effort

The goal isn’t just resolving queries more efficiently, it’s creating capacity to do higher-value work elsewhere.

These are your fast paths to ROI. Some AI Agents, like Fin, can resolve a large percentage of your support volume straight out of the box. (The average Fin customer sees 65%.)

Once you’ve identified a set of high-value, quick-win opportunities, you’ll have the foundation for an AI roadmap that connects your strategy directly to business impact.

This is your starting line. From here, you can begin testing AI against these use cases and proving value quickly.


2. Prove the basic economics

AI changes the economics of customer service. It breaks the linear relationship between business and headcount growth, unlocking the ability to scale without the traditional constraints of hiring.

In this new model, your team can handle more conversations without adding more people, which breaks the logic of traditional support economics.

What traditional support economics miss

Legacy models typically focus on a single metric: dividing total support cost by the number of conversations (or cases) handled. But this is an oversimplification. Here’s why:

  • Conversation types varySome queries are simple, others complex and time-intensive.
  • Human resolution rate isn’t 100%It’s easy to assume that human support resolves every customer issue, but just because a conversation was handled doesn’t mean it was fully resolved.
  • Team costs go beyond salariesOverhead, tools, training, and attrition all add up.
  • Automation value depends on qualityIf your AI Agent doesn’t resolve queries effectively, it doesn’t save money.

In short, surface-level metrics like average cost per conversation don’t capture the full picture, because not all conversations carry equal weight, and not all costs are obvious. A cheap or “free” AI solution might look efficient on paper, but if it doesn’t actually resolve conversations, those interactions still flow through to your human team – bringing all of the associated costs back into play.

To understand the true economics of AI, you need a more nuanced model that reflects the actual outcomes, distribution of effort, and real cost of human work.

How to measure the real value of AI support

Operating in an AI-first world means you have to think about value differently:

The work you automate determines the value

Not all support work is equal, and neither is the value of automating it.

Solving a high volume of easy questions can reduce some costs, but the real payoff comes when you start automating the complex, time-consuming conversations that tie up your team and slow down resolution.

That’s where AI Agents deliver asymmetric value. They’re not just built for deflecting FAQs. They’re designed to handle complex, nuanced conversations that require real understanding, judgment, and context. The more you trust it with meaningful work, the more your return compounds.

At a certain point, when your AI Agent reaches a meaningful automation rate (AI Agent involvement rate × AI Agent resolution rate) you cross the breakeven threshold. Every resolution beyond that point is a high-yield investment.

But ROI isn’t linear. Automating 10% of conversations doesn’t mean you cut 10% of costs. Work isn’t evenly distributed, and time savings often come in fragmented chunks across shifts and teams, so it rarely maps neatly to headcount reduction. But that doesn’t diminish the value. AI doesn’t need to replace full-time roles one-for-one. It frees up capacity, absorbs repetitive effort, and enables you to reallocate time. The more you automate, the more efficient, scalable, and cost-effective your support model becomes. The ROI doesn’t level off, it accelerates.

Understand your fully loaded cost per conversation

You don’t need to model every line item to understand the economics. But you should recognize that your true human cost per conversation goes far beyond salaries:

  • Benefits and taxes.
  • Software and equipment.
  • Management overhead.
  • Coaching, training, and QA.
  • Attrition, backfill, and onboarding.

Factor all of this in and you’ll get a more realistic cost baseline.

Note: Industry benchmarks put the average cost per conversation well north of $1, and often much higher for complex queries. That’s your breakeven point. Once your AI Agent starts resolving conversations below that threshold, it’s not just cost-neutral, it’s ROI-positive.

This is why performance is such an important component to understanding the value of an AI Agent. If it doesn’t resolve, it doesn’t return.

Look at outcomes, not just unit cost

A cheaper AI Agent isn’t better if it resolves fewer conversations. How often the AI actually solves the issue is what matters.

A low-cost, lower-performing AI Agent might seem like a better deal at first, but if it can’t resolve conversations effectively, it ends up costing you more. More escalations. More human handoffs. Less efficiency. Worse customer experience.

For example, an AI Agent that costs $1 per resolution and resolves 50% of conversations is far more valuable than one that costs $0 per resolution but only resolves 30%, because unresolved conversations still require human follow-up, which costs time, money, and even customer satisfaction.

You also need to consider the total cost of ownership – onboarding, integration, training, and maintenance all affect the true ROI.

A simple formula to calculate cost savings

Once you’ve accounted for the variables above, you can use this high-level formula to estimate your potential savings:

((HUMAN COST PER RESOLUTION) - (AI COST PER RESOLUTION)) X (NUMBER OF AI RESOLUTIONS)

Here’s an example of what this might look like:

VariablesCompany XCompany Y
Monthly Conversation Volume5050
Resolution Rate50% (average resolution rate)30% (assumption)
Number of AI Resolutions2515
Price per AI Resolution$1$0
Total AI Price$25$0
Number of Human Resolutions2535
Price per Human Resolution$15$15
Human Support Cost$375$525
Total Cost$400$525

This is just a starting point, but it gives you a clear way to model potential economic impact and begin forecasting the value of AI in your support organization.

If you want to get a sense of what this could look like specifically for you, use our ROI calculator.

Value beyond cost savings

Cost savings are only one aspect of AI-support economics. In early phases, AI unlocks clear efficiency gains. But the real transformation happens when AI starts to shift the entire shape and purpose of your support model.

Without AI, your team will eventually hit structural limitations:

  • Scalability stallsYou can’t grow support at the pace of your business.
  • Budgets inflateHeadcount costs rise linearly.
  • Customer expectations outpace your teamCustomers expect instant, AI-enabled support – and churn if they don’t get it.
  • Your agents burn outRepetitive tasks drive attrition.

When you adopt AI fully, however, the equation flips:

  • You decouple support growth from headcount growth.
  • You reinvest human time where it matters most.
  • You redesign support as a strategic lever for growth.

This is where AI starts to pay compounding returns. The real gains don’t come from cost cutting alone – they come from reallocating time and expanding the types of value AI can deliver directly.

As AI becomes more adept at answering complex questions and taking meaningful actions, you’ll need to rethink the economics of delivering customer service – even down to how you define productivity across humans and AI Agents.

With AI handling the bulk of your support volume, both AI and human teammates can focus on higher-value efforts, such as:

  • Driving deeper product adoption.
  • Increasing retention and LTV.
  • Turning support into a value driver.

This shifts the cost/value equation in a meaningful way:

  • Higher cost per human conversation.
  • Lower total cost per resolution overall.
  • Much higher value per interaction – whether delivered by a human or an AI Agent.

But that’s not a bad thing – it’s a sign that the work your support team is doing has evolved.

AI enables scale and value. Humans refine the system and focus on high-impact moments. Together, they make support a driver of growth, not just a cost to manage.


3. Decide whether to build or buy an AI Agent

In most cases, you shouldn’t build your own AI Agent.

It’s getting increasingly easier to build AI-powered software, which makes it tempting. But if you do it for the wrong reasons or misjudge the trade-offs, you’ll likely waste time and money.

Intercom

I get the appeal of building your own system – it’s fun, it’s a great learning experience, and there’s something special about shipping your own AI code. But my advice? Channel that energy into your own product, not a non-strategic side quest.

DARRAGH CURRANCTO AT INTERCOM
Darragh Curran

You should only consider building if:

  • You have a specialized AI team.
  • You operate at massive scale.
  • You have extremely niche support needs.

If not, you’re better off buying. And even if you do have all of those things, building is still not always the best option.

While DIY may look simple and cheap at first, building an AI Agent is rarely just a matter of connecting an LLM to a UI. It quickly turns into a big infrastructure investment with hidden development, product management, integration, and upkeep costs, unreliable performance, and limited strategic upside.

Building can sound appealing because:

  • You think it gives you full control.
  • You believe it’ll be cheaper over time.
  • You want something tailored to your use case.

But that control comes with a heavy operational burden. It means building and maintaining a full system. You’ll need:

  • A retrieval-augmented generation (RAG) pipeline for high-quality responses.
  • Query rewriting and content ranking to structure ambiguous customer questions.
  • LLM orchestration logic across vendors and model versions.
  • Confidence scoring, handoff frameworks, fallback logic, and human-in-the-loop guardrails.
  • A process for feedback loops, versioning, regression testing, and ongoing performance improvement.
  • A team of the right people who are keeping up with the technology, moving fast, and adapting.

The bar to beat best-in-class solutions is extremely high

If you go the build route, you’re not competing with today’s market. You’re competing with where top AI Agents will be six months from now. You’ll need to keep pace on quality, safety, reliability, and coverage, without the benefit of the ongoing innovation and infrastructure investment provided by a shared platform.

That’s a serious commitment, and for most companies, it’s one that doesn’t align with their core focus or resourcing.

In most cases, the time, cost, and risk just don’t pencil out.

If your goal is speed to value, high performance, and ongoing innovation, buying an AI Agent is the faster, safer path. You get:

  • Faster time to deployment and ROI.
  • Ongoing improvements built on usage data from multiple customers.
  • Reduced risk around performance, security, and maintenance.
  • The ability to focus your teams on higher-leverage work: improving content, training the AI Agent, and analyzing impact.

If you have a strong AI team, massive support volume, and very specialized needs, it might make sense to build your own AI Agent. But even Anthropic, one of the leading AI labs, uses our AI Agent Fin because they recognize the constant iteration required to stay safe, accurate, and deeply integrated with support workflows.

For most companies, buying is the smarter path. You’re not just buying technology. You’re buying time, trust, and future flexibility.

Anthropic

If you're debating whether to build your own AI solution or buy one, as a fast-growing company in a complex space, my advice would be to buy – and specifically, buy Fin.

ISABEL LARROWProduct support operations at anthropic
ISABEL LARROW
2.4

Step 2:
Evaluate the AI Agent

Define your criteria for selecting
and evaluating an AI Agent

When choosing and evaluating AI Agents, there are a number of factors you need to consider. You need to understand if it can integrate with your business, handle complexity, operate at scale, perform well, and deliver value.

We recommend breaking your assessment into two parts:

ENTRY CRITERIAUse these to assess whether an AI Agent is viable for your business, technically, securely, and operationally.
CapabilitiesQuestion to answer: Does the AI Agent offer advanced capabilities, i.e., can it handle more complex, multi-turn resolution processes, or provide continuous improvement insights?
Platform fitQuestion to answer: Can this AI Agent actually work in your environment, technically, securely, and operationally?
Evaluation criteriaOnce you’ve chosen your AI Agent to test, use these to measure performance.
Business performanceQuestion to answer: Does the AI Agent actually resolve the kinds of conversations your team handles today?
Conversation qualityQuestion to answer: Can the AI Agent hold up in real conversations with your customers?

Entry criteria


1. Capabilities

Does the AI Agent support the use cases you care about most?

Look for:

  • Task automation, not just answering FAQ-style questions.
  • Personalization using customer data.
  • Ability to guide and control the behavior of the AI Agent.
  • Multi-language, multi-channel support.
  • AI insights for real-time customer experience analysis.
  • Seamless handovers to human agents.
  • Transparency and control over the end-to-end experience.

Why it matters: Capabilities only create value if they lead to real outcomes. The most important measure of any AI Agent is its ability to fully resolve conversations – consistently, accurately, and at scale. That's the foundation for ROI, customer trust, and long-term success. Just as critical is transparency and control: teams need to understand how the AI Agent makes decisions, guide its behavior, and ensure it's operating in line with brand, policy, and quality standards. Without this, even advanced features can introduce risk and undermine trust.


2. Platform fit

Determine if the AI Agent can integrate with your existing systems, meet your compliance standards, and scale with your team.

Areas to focus on here are:

Integrations and customization

Can it connect with your existing systems? How easily can it be customized to fit your workflows?

Check that:

  • It connects with your helpdesk, knowledge base, CRM, and analytics tools.
  • It's extensible via APIs, webhooks, or SDKs.
  • It can reflect your business logic, data sources, and routing rules.
Security and compliance

Does the vendor meet your data privacy obligations?

Confirm that:

  • The vendor meets privacy and data protection obligations (e.g., GDPR, CCPA, HIPAA.).
  • Certifications like SOC 2, ISO 27001, or ISO/IEC 42001 are in place.
  • It can isolate or redact PII.
  • It can support SSO, RBAC, and audit logs.

Why it matters: The AI Agent you choose needs to meet your company's standards. Platform fit ensures a safe, scalable deployment.

Evaluation criteria


1. Business performance

This is where operational value becomes real. Beyond conversation quality, you need to measure the actual results the AI Agent is producing for your business and customers.

Some metrics to measure:

  • Resolution rateConversations fully handled by the AI Agent with no human involvement.
  • Deflection or containment rateConversations handled without reaching your team.
  • Time savedHours of manual work offloaded from your support team.
  • CSATIf you're testing in a live environment, track how comparable customer satisfaction with AI is to human-handled interactions.

When it comes to measuring AI's impact, we recommend:

1. Focusing less on deflection, and more on resolution

While helpful as an early signal, deflection is a limited, and potentially misleading, measure of success. What matters is whether the issue was resolved. With AI Agents like Fin now capable of resolving the majority of queries, we need to shift our focus from deflection to outcomes.

2. Measuring customer satisfaction across all conversations with AI

Across all Intercom customers who use CSAT, just 8% of conversations get scored, whereas AI can analyze 100% of customer conversations (involving both AI Agents and humans) in real time. If you're evaluating Fin, it assesses three critical dimensions of customer conversations:

  • Resolution statusWas the issue actually resolved? And if there were multiple issues, were each of them resolved?
  • Customer sentimentHow did the customer feel throughout the interaction?
  • Service qualityWas the response clear, helpful, and efficient?
Fin uses these inputs to generate a "CX rating" from 1–5 for each conversation. These individual ratings contribute to a broader "Customer Experience Score," based on real-time insights from every support conversation.
CX Score

Why it matters: Business value comes from resolving the right volume, reliably, securely, and at scale.


2. Conversation quality

Quality isn't just about how good the AI Agent sounds. It's about how clearly and accurately it helps customers.

You need to look at:

  • AccuracyDoes it understand intent and deliver the right answer, every time?
  • BehaviorDoes it know when to clarify or hand over to a human agent? Can it represent your brand's tone of voice and adhere to your company's policies?
  • ExperienceDoes it create a smooth, fast, and satisfying experience for both customers and support teams?

Why it matters: Great AI Agents build trust by communicating clearly, responding appropriately and accurately, and staying on brand. Incorrect or vague answers, poor tone, or robotic behavior can undermine confidence.

How to run an evaluation


This four-step process helps you apply both evaluation lenses – business performance and conversation quality – in a structured, outcome-driven way.

Note: You don't need to run a multi-vendor evaluation to make a confident decision. In many cases, a single-threaded proof of concept (POC) with the strongest-fit solution is the fastest and clearest path forward. It gives you more control, lets you go deeper, and builds a stronger signal around how the AI Agent performs in your real-world environment.

Step 1: Define what success looks like

Before testing an AI Agent, align your success criteria and metrics to what matters most across the two evaluation lenses – business performance and conversation quality.

Use a mix of quantitative and qualitative metrics to get a complete view of value and make a compelling case for adoption.

For example:

Business performance
Use quantitative metrics like:
  • Resolution rate
  • Deflection/containment rate
  • Time saved
  • CSAT (if testing in a live environment)
  • Customer Experience Score (if testing in a live environment)
These will provide measurable proof of the AI Agent's impact, making it easier to justify investment and compare against benchmarks.
Conversation quality
Use qualitative signals like:
  • It understood what the customer was asking.
  • It answered questions accurately and on-brand.
  • It knew when to hand over to a human agent and did it smoothly.
These capture subjective but crucial elements like trust, usability, and perceived value – all of which are key drivers in decision-making.

Note: We do not recommend tracking metrics in isolation. By combining both quantitative and qualitative metrics, you'll get a more complete view of the AI Agent's impact. This approach will also help ensure the evaluation isn't just about hitting numbers, but also about demonstrating real-world fit and usability, addressing both executive and customer concerns.

Clay

Keep what’s important to your business up front and center. For us, that was transparency and control over the customer experience. Focusing on the end goal helped us come to the right decision because we knew what was important to us

George DiltheyHead of Support
George Dilthey

Step 2: Build a realistic test environment and train the AI Agents

Once you've defined what success looks like, you can begin testing.


1. Set up your test environment using real customer questions and your current knowledge base or help content

You can choose to test in a sandbox environment or with live conversations. The important thing is to use real customer questions to test the AI Agent from a business performance and conversation quality perspective.

Once you are confident with the baseline performance, we recommend testing the AI Agent with real users to validate the quantitative and qualitative metrics in the real world.

Source a range of customer conversations to test against:

  • Complex queries that typically require multiple touchpoints from different team members.
  • Vague queries that don't contain any "real" information and require further clarification to resolve.
  • Edge cases that have been difficult for your human team to resolve.
  • A few sensitive scenarios, such as billing disputes and cases where customers have become frustrated.
  • Examples of queries in different languages, if you provide multilingual support.

Take this a step further and prepare variations of the same questions to test how the AI Agent handles different types of communication:

  • Difficult questions that require information from multiple sources to answer.
  • Different phrasings of the same question.
  • Incomplete or fragmented queries.
  • Questions with typos or grammatical errors.
  • Conversations with various levels of formality.

The goal is to simulate what happens in reality. Any AI Agent can look impressive in a controlled setting, but performing well when faced with real challenges customers bring is what separates "good enough" from great.

If you're evaluating more than one solution, make sure you set up your AI Agents in the same way for a fair comparison. Split out the conversation volume equally so you can accurately test each of the solutions.


2. Train the AI Agent

Prepare your knowledge base or help content

You need quality content for an AI Agent to deliver good results. Assess your knowledge base content for:

  • CoverageMake sure you have adequate coverage for the testing cohort to give the AI Agent all the information it needs to address key questions and topics. For example, if you want to test whether the AI Agent can fully resolve queries on a specific topic, like your accounting product, or for a specific audience, like your Freemium users, it must have access to relevant content for both.
  • AccuracyTo prevent the AI Agent from learning outdated information, make sure what you're exposing it to is accurate and up to date. For example, if your return policy has changed from 60 days to 30 days, update this.
  • StructureThe more straightforward and comprehensive your articles are, the easier it will be for the AI Agent to consume them. Focus on simple language and an easy-to-scan structure.

You don't have to reformat or rewrite all your help content before running tests. This is just something to be aware of and potentially return to should content gaps emerge during the testing or you spot any glaring issues.

Configure the AI Agent's rules, tone, and behavior

Modern AI Agents let you control how they communicate and act – for example, you can instruct them to provide concise or comprehensive responses, use specific terminology for your industry, or follow protocols that match your support policies.

For Fin, we call this Guidance. It enables you to define Fin's communication style, coach it to gather context and clarify issues, and set rules for routing and handovers.

Guidance

Set these parameters in advance to ensure you're testing the AI Agent in a way that reflects your brand, policies, and expectations. This will give you a more accurate view of how it will perform in your real-world environment.

Step 3: Score performance and analyze results

Run your test conversations through the AI Agent and evaluate results through your two performance lenses:

  • Business performanceHow well does it deliver results?
  • Conversation qualityHow well does it communicate?

Business performance

Resolution rate
How many queries did the AI Agent resolve end-to-end?
    Deflection/containment rate
    How many queries did the AI Agent manage without needing to hand it over to a human?
      Time saved
      How quickly did the AI Agent resolve queries, and how much time would it have taken your team to handle them?
        CSAT* (if in a live environment)
        • How are customers rating the AI Agent vs human agents?
        • [If you're testing Fin] What's the Customer Experience (CX) Score?

        Conversation quality

        Accuracy
        • Did the AI Agent understand the customer's intent? Did it ask for clarification when needed?
        • Did it pull from the right knowledge sources?
        • Did it personalize the response appropriately for the customer's context?
        Behavior
        • Did it maintain the right tone for your brand?
        • Did it escalate appropriately when issues were beyond its scope?
        • Did it route to the correct team when handoffs were needed?
        Experience
        • Was the overall interaction smooth and satisfying?
        • How would a typical customer rate this response?
        • Would your support team feel confident standing behind this answer?

        Important note: you should also evaluate the vendor, not just the AI Agent

        The vendor behind the AI Agents matters just as much as the solution itself. You're choosing a partner for transformation. One that will help you evolve how your business delivers customer experience.

        This isn't traditional vendor management. You're betting on a vision of the future. Ask:

        Are they pushing boundaries?

        • Are they shaping the future of AI-powered customer experience, or reacting to it?
        • Do they have a clear point of view on where AI is headed?
        • Are they building capabilities that stretch beyond today's benchmarks, and not just keeping pace?

        Are they a true partner, not just a provider?

        • What does their product roadmap look like? How does customer feedback shape it?
        • What kind of support will you get post-launch, e.g., ongoing support, or does the relationship shift to basic technical support?
        • Are they transparent about current limitations? Vendors who acknowledge gaps and commit to fixing them are more likely to be honest partners than those who oversell capabilities.

        Are they built for long-term success?

        • How long do companies like yours typically stay with this vendor? Look at their existing customer base, retention, and growth rates.
        • How do they respond to hard questions during evaluation? Those who get defensive about limitations or rush you toward a decision may not be committed to long-term success. The best vendors welcome hard questions and help you spot risks before they become problems.

        Ask yourself: does this vendor feel like someone who will help us reinvent customer experience, or just someone selling software? Great AI Agents are backed by great partners. Look for vendors that are obsessed with support, transparent about how the technology works, and committed to co-building the future with you.

        Clay

        We're in a world where lots is changing and we're still learning about AI. The most important thing I would express to other CX leaders is that it's crucial to have the right partners alongside you to teach you what you need to know to be successful in this world.

        Jess BergsonHead of CX
        Jess Bergson

        Step 4: Decide whether the AI Agent is the right fit

        Weight your findings based on your personal priorities. If accuracy is non-negotiable, don't compromise, even if other qualities like tone or personality feel strong. If you need immediate deployment, factor in integration complexity and vendor support quality.

        Consider both immediate performance needs and long-term operational success.

        The trade-offs you have to consider here are:

        • If business performance is low, you'll struggle to show ROI.
        • If conversation quality is bad, customer trust may suffer.

        Ultimately, the AI Agent you choose should be the one that fits your goals, supports your team, and will help you scale sustainably.

        2.5

        Deploy your AI Agent

        Once you've chosen your AI Agent, the next step is to deploy it.

        If you didn't deploy the AI Agent to real customers as part of your initial evaluation, now is time to start rolling it out – integrating it with your support channels, letting it handle real conversations with your customers, and seeing it deliver meaningful value.

        Build a deployment plan

        Start by translating your business case and evaluation results into a deployment roadmap. This keeps your rollout focused, scoped, and tied to the goals you've already identified.

        Here's what you need to define:

        Deployment scope

        Start by identifying the use cases you want your AI Agent to handle. Map out the ones you'll want to address now, next, and later. This gives you a roadmap to full automation, but also keeps things grounded by outlining the initial deployment targets – the conversations your AI will handle first. (You can use the use cases identified in your business case to get started, but now's the time to think beyond that.)

        Rollout model

        There are two possible approaches here:

        • Fast-trackLaunch to a broad audience and iterate in production.
        • Phased rolloutStart with specific topics, user segments or channels and expand in waves.

        What you choose will be dependent on your company, industry, and customers.

        Here's a way to think about this:

        ROLLOUT STYLEBEST FORTRADE-OFFS
        FAST-TRACKStartups, fast-moving teams, bold exec sponsorsFastest feedback and ROI. Higher stakes. Requires strong alignment.
        PHASEDEnterprises, regulated industriesDe-risks launch. Slower path to impact.

        If you choose to phase, here's a model for expanding gradually:

        • Phase 1High-volume topics, low-risk customer segments, specific channels.
        • Phase 2Related topics, broader customer groups, new channels.
        • Phase 3More complex scenarios, edge cases, all customers, all channels.

        Prepare your team

        A successful deployment requires clear accountability. Assign owners for:

        • Content qualityEnsuring your AI Agent has the right information, in the right format.
        • System configurationManaging integrations, workflows, and behavioral tuning.
        • Performance trackingMonitoring early results and identifying opportunities to optimize.

        Determine how you'll measure success

        Set checkpoints.

        If your target is a 65% resolution rate within 90 days, what does success look like at week 2, week 4, week 8?

        Prepare your AI Agent for production

        Your AI Agent is only as strong as the content it's trained on and the configuration behind it.

        Prioritize four areas of readiness:

        • Content

          Knowledge your AI Agent has access to.

          Use this comprehensive Knowledge Management Guide if you're looking for detailed guidance.

        • DataConnect relevant data sources, like your CRM or billing platforms, so your AI Agent can deliver contextual, customer-specific responses.
        • BehaviorSet behavioral rules and determine its tone of voice. This ensures your AI Agent speaks with the same clarity and consistency as your best human agents.
        • OrchestrationConfigure the high-level settings for the AI Agent – what languages it speaks, what channels it's available on, and how it handover to a human agent when required.

        Launch to customers

        Before the AI Agent goes live, test the full customer experience:

        Ask:

        • Is the AI Agent available in the right channels?
        • Are customer segments clearly defined?
        • Is the introduction message clear and on-brand?
        • Are handover flows between humans and AI Agents easy and reliable?
        • Are feedback loops (e.g. CSAT surveys) in place?

        When you're happy with the experience, deploy your AI Agent and begin monitoring performance. If you're approaching the launch in phases, use this step to launch Phase 1, validate assumptions, learn fast, and build momentum.

        Apply your success framework

        Now that you're live, use the success metrics you previously defined to track real-world performance, for example:

        • Resolution rateAre you hitting your benchmark for end-to-end automation?
        • Customer satisfactionIs CSAT or your Customer Experience Score steady or improving?
        • Efficiency gainsAre you saving time and reducing manual volume?

        If the results are off-track, don't panic. Most early issues stem from content gaps, configuration issues, or data access. This becomes the next part of the process: moving from deployment to always-on monitoring and optimization. We'll break that down in the next chapter.

        3

        What's next?