Simulation Testing

Overview

Simulation testing is a powerful step in your agent development process that enables scalable, automated validation of conversational flows. It helps catch complex scenario gaps or unexpected behaviors that manual tests might miss, so you can confidently iterate and deploy your AI agent.

Use Test Audio and Test Logic during the agent-building phase to verify each component of your agent works as intended. Then leverage simulation testing before go-live to catch potentially missed issues across multiple scenarios or after live iterations to ensure expected behaviors remain consistent.

✅ What you can test with simulation testing

  • Conversational logic: Verify the agent follows task flow correctly under diverse scenarios, handles objections/questions, and integrates RAG (retrieval-augmented generation) responses.
  • Custom actions triggered during conversation.
  • Simultaneous testing of multiple scenarios using the same agent prompt.

❌ What you cannot test with simulation testing

  • Sound, speed, volume of the voice
  • Responsiveness and interruption handling of your agent
  • Call transfers or E2E latency over telephony
  • Personalization using test values for dynamic variables (coming soon)
  • Personalization from actual profile data
  • Dispositions (coming soon)

Create Test Cases

  • NameA brief identifier summarizing what the test case covers
  • Contact PromptA descriptive input prompt that defines the contact’s behavior and dialogue during simulation, including how the contact responds and relevant behavioral traits.
    • Best practice: Treat this like a mini prompt. Use ## to separate sections (similar to agent prompts).
    • Include a Custom Scenario describing what the contact says in your test scenario (most important).
    • Optionally add Customer Persona and Customer Attributes to add personality or traits.
    • Note: This section does not populate personalization variables in the agent prompt; it only shapes contact behavior in the simulation.
  • Success CriteriaA reference defining the expected agent outcomes. Deviations may indicate performance gaps. (Auto-evaluation coming soon.)

Auto-generate Test Cases

Click Get Me Startedto auto-generate a list of 10 test cases based on existing agent prompt.This option is available only when no test cases exist yet.

You can edit generated test cases to better suit your needs.

Simulate Tests

Select the test cases to run simulations. Run individually or in bulk.


  • During simulation, the outcome column will show In Progress until the conversation concludes.
  • Click the test case name to view transcripts, knowledge base retrievals, and function calls.
  • Review transcripts carefully to identify next steps.

Iterate Test Cases

At your first simulation round, the contact may behave unexpectedly (e.g., long paragraphs instead of succinct answers). Focus first on refining test cases by editing the contact prompt and rerunning simulations. This upfront effort is critical — once test cases are reliable, you can run them repeatedly without manual retesting.

What's important to cover?

  • End-to-end Task Steps: If your agent has multiple task steps with branching logic, create enough test cases to cover the key paths.
  • Objections & Questions: Include varied objection handling and common questions.
  • Guardrails: Test agent adherence to defined guardrails and script boundaries.
  • Edge Cases: Add stress tests with odd or unexpected customer behaviors.

Iterate Agent

After finalizing test cases, observe the agent’s behavior for:

  • Consistent triggering of expected functions, even in complex conversations.
  • Accurate and prompt responses to multiple customer questions.
  • Effective retrieval and use of relevant knowledge base content.
  • Adherence to guardrails and agent style guidelines.

Coming Soon

  • Auto-eval against success criteria with pass/fail
  • Cloning test cases
  • Have feedback? We'd love to hear from you - reach out to us with your thoughts!