العودة إلى المدونة
Auto-EscalationUXSupport

Auto-Escalation: How AI Chatbots Should Handle "I Don't Know"

LaunchChat Team6 min read

The "I Don't Know" Problem

Every AI chatbot has limits. No matter how good your documentation is, users will ask questions the chatbot can't answer:

  • Questions about topics not in your docs
  • Highly specific edge cases
  • Account-specific issues ("why was I charged twice?")
  • Bug reports that need human investigation
  • Emotional situations that need empathy

The question isn't whether your chatbot will fail — it's what happens when it does.

The Wrong Way: Silent Failure

Most chatbots handle uncertainty poorly:

Generic refusal: "I'm sorry, I can't help with that. Please contact support."

This is the worst possible response because:

  1. It doesn't acknowledge what the user asked
  2. It doesn't explain why it can't help
  3. It forces the user to start over with a human agent
  4. The user has to re-explain their problem
  5. Nobody learns that this question was asked

The user experience goes from "instant AI help" to "dead end" in one message.

Auto-escalation flow
Auto-escalation flow

The Right Way: Graceful Escalation

Auto-escalation is the process of smoothly transitioning from AI to human support when the AI can't help. Done well, it feels natural and preserves the user's context.

How LaunchChat Handles It

LaunchChat's auto-escalation works in three stages:

#### Stage 1: Confidence Assessment

Every answer gets a confidence score based on:

  • Retrieval similarity: How closely the retrieved chunks match the question
  • Coverage: Whether the chunks contain enough information to answer fully
  • Specificity: Whether the answer addresses the specific question asked

You configure a confidence threshold per widget (default: 0.6). Below this threshold, the AI doesn't attempt an answer.

#### Stage 2: Honest Refusal

When confidence is below threshold, the chatbot responds with a configurable refusal message. The default:

"I don't have enough information in my knowledge base to answer that accurately. Let me connect you with someone who can help."

Key elements:

  • Honest: Admits it doesn't know (not "I can't help")
  • Specific: References the knowledge base (not a vague "I'm sorry")
  • Forward-looking: Offers a next step (not a dead end)

#### Stage 3: Escalation Form

After the refusal, the widget shows an escalation form:

  • User's name
  • Email address
  • The original question (pre-filled)
  • Optional additional context

The form submission creates an escalation record that includes:

  • The full conversation history
  • The question that triggered escalation
  • The closest matching chunks (so the human agent has context)
  • Timestamp and widget metadata

What the Human Agent Sees

When an escalation comes in, the agent gets:

  1. The user's question
  2. The full conversation (what the user asked before)
  3. What the AI tried to answer (if anything)
  4. The closest matching documentation (even if below threshold)
  5. Contact information

This context means the agent doesn't start from zero. They know what the user asked, what the AI found, and where the documentation falls short.

Configuring Escalation

Confidence Threshold

The threshold controls how strict the AI is:

  • 0.8+ (strict): AI only answers when very confident. More escalations, but fewer wrong answers.
  • 0.6 (default): Balanced. AI answers most questions but escalates edge cases.
  • 0.4 (permissive): AI attempts more answers. Fewer escalations, but some answers may be less accurate.

For support chatbots, 0.6 is usually right. You want the AI to be helpful but not make things up.

Refusal Message

Customize the message to match your brand voice:

  • Formal: "I wasn't able to find a definitive answer in our documentation. Would you like to speak with our support team?"
  • Casual: "Hmm, I'm not sure about that one. Want me to connect you with a human?"
  • Technical: "This question falls outside my current knowledge base. I can escalate to our engineering team."

Escalation Destination

Where do escalations go?

  • Email: Sent to your support email address
  • Dashboard: Visible in the LaunchChat activity feed
  • Both: Email notification + dashboard record (recommended)

The Knowledge Gap Connection

Auto-escalation and knowledge gap tracking work together:

  1. User asks a question → AI can't answer → escalation triggered
  2. The question is logged as a knowledge gap
  3. You see the gap in your dashboard with frequency data
  4. You write the missing documentation
  5. Next user with the same question gets an AI answer
  6. No more escalations for that topic

This feedback loop means your escalation rate decreases over time. Typical trajectory:

  • Week 1: 30-40% escalation rate
  • Month 1: 20-25% after filling top gaps
  • Month 3: 10-15% after systematic gap-filling
  • Month 6: 5-10% (mostly account-specific issues)

Anti-Patterns to Avoid

Don't Hallucinate

The worst thing a chatbot can do is make up an answer when it doesn't know. This is worse than saying "I don't know" because:

  • Wrong answers erode trust
  • Users may act on incorrect information
  • It's harder to recover from a wrong answer than a non-answer

LaunchChat's confidence threshold prevents this. Below the threshold, the AI refuses rather than guesses.

Don't Loop

Some chatbots ask clarifying questions when they can't answer, hoping the user will rephrase. This creates frustrating loops:

User: "How do I cancel?"

Bot: "Could you be more specific about what you'd like to cancel?"

User: "My subscription"

Bot: "I'm not sure I understand. Could you rephrase?"

If the AI doesn't have the information, no amount of rephrasing will help. Escalate promptly.

Don't Hide the Human Option

Some chatbots make it difficult to reach a human, forcing users through multiple AI interactions first. This is hostile UX. If the user wants a human, let them reach one.

LaunchChat always shows an escalation option after a refusal. Users can also manually trigger escalation at any point in the conversation.

Don't Lose Context

When escalating, pass the full conversation to the human agent. Nothing is more frustrating than explaining your problem to an AI, getting escalated, and having to explain it all over again.

Measuring Escalation Quality

Track these metrics:

  • Escalation rate: % of conversations that escalate (target: under 20%)
  • False escalations: Conversations that escalated but could have been answered (indicates threshold is too high)
  • Missed escalations: Conversations where the AI answered incorrectly instead of escalating (indicates threshold is too low)
  • Resolution after escalation: How quickly human agents resolve escalated issues

Getting Started

Auto-escalation is enabled by default in LaunchChat. To configure:

  1. Go to Widget Settings → Behavior
  2. Set your confidence threshold (start with 0.6)
  3. Customize the refusal message
  4. Set your escalation email
  5. Test by asking questions not in your docs

The goal is a chatbot that's helpful when it can be and honest when it can't. Auto-escalation makes that possible.

Try LaunchChat free — auto-escalation included on all plans.