Frequently Asked Questions
1. How do credits work in Cekura, including pricing for testing, monitoring, and overages?
Understanding Credits and Pricing in Cekura
Cekura uses a credit-based system to manage usage across testing, monitoring (observability), and evaluation. This allows you to pay only for what you use across different stages of your agent's lifecycle.
1. Credit Consumption Rates
Credits are consumed based on the specific activity performed within the platform:
- Voice Testing: 5 credits per minute of voice run.
- Chat-Based Testing: 0.5 credits per message sent by the testing agent.
- Monitoring & Observability (Evaluation): 0.2 credits per metric run to evaluate a conversation.
2. Usage Examples
Example A: Testing a Voice Agent
If you run a 2-minute voice test and evaluate it with 5 metrics, the total cost is 11 credits:
* Voice Duration: 2 minutes × 5 credits = 10 credits
* Evaluation: 5 metrics × 0.2 credits = 1 credit
Example B: Monitoring/Observability (Call Evaluation Only)
If you are importing external calls (e.g., from Retell or Twilio) for quality assurance and run 10 metrics to evaluate the performance of that call, the total cost is 2 credits:
* Evaluation: 10 metrics × 0.2 credits = 2 credits
3. Overages and Manual Top-ups
To ensure your testing and monitoring are never interrupted, Cekura provides flexible credit management options:
- Overages: When enabled, overages allow the platform to continue processing calls and metrics even after your base credit limit is reached. This is critical for production monitoring to ensure you don't lose data during high-volume periods. Overages are typically billed at your standard unit rate.
- Manual Top-ups: If you are in a sandbox or trial phase and run out of credits, our support team can manually add credits to your account to ensure your development work continues without delay.
4. Best Practices for Credit Optimization
To manage your credit burn effectively, consider the following strategies:
- Smart Sampling: For high-volume monitoring, configure your metrics to run only on a percentage of calls or specific call types (e.g., excluding calls that go to voicemail) to save on evaluation credits.
- Bulk Re-evaluation: Use the re-evaluate feature carefully. Note that re-running metrics on existing calls will consume credits at the standard rate of 0.2 credits per metric. This is often used when you have updated a metric prompt and want to see how it performs on historical data.
- Metric Refinement: Periodically review your evaluators. Disabling metrics that are no longer providing actionable insights will reduce your per-call credit cost.
- Use the Optimizer: For project-level metrics, use the Cekura Optimizer to ensure your prompts are efficient, which helps maintain high accuracy without unnecessary manual re-runs.
If you need to request a credit top-up, enable overages, or discuss a custom plan, please reach out to us at support@cekura.ai or via your dedicated Slack support channel.
2. Where can I upload knowledge base files on the Cekura dashboard and what is the recommended format for an FAQ-based document?
To provide your AI agent with the necessary context and reduce hallucinations, you can upload knowledge base files directly through the Cekura dashboard.
How to Upload Knowledge Base Files
- Log in to your Cekura dashboard.
- Navigate to Agent Settings for the specific agent you are configuring.
- Look for the Knowledge Base section, typically located on the bottom right of the settings page.
- Click to upload your file.
Recommended Formats and Content
Cekura supports a variety of file formats for your knowledge base, including:
- Text (.txt)
- PDF (.pdf)
- CSV (.csv)
- JSON (.json)
Best Practices for Content:
While you can upload transcripts of calls manually tagged as hallucinations to help the system learn, we primarily recommend providing a structured FAQ-based document. This format is highly effective for grounding the agent and ensuring it has clear, factual references to pull from during interactions.
By default, Cekura defines a hallucination as an instance where the agent provides information that was not present in its prompt or its uploaded knowledge base. Providing a comprehensive FAQ document is the most reliable way to prevent these occurrences.
For more detailed information on setting up your agent's knowledge base, please refer to our Agent Setup Guide.
3. If I add a custom metric to a specific agent, will it automatically apply to all agents or can it be applied to a single agent only?
In Cekura, you can choose whether a metric applies to a single agent or across your entire project. By default, metrics are created at the Agent-level, meaning they only apply to the specific agent they were created for.
Agent vs. Project Metrics
- Agent-level Metrics: These are specific to one agent and are ideal for unique evaluation criteria tailored to a specific persona or use case.
- Project-level Metrics: These apply to all agents within a project, allowing for consistent evaluation across your entire workspace.
How to Manage and Create Metrics
- Navigate to Metrics: Select the Metrics section from the left-hand navigation panel.
- Switch Views: Use the tabs at the top of the page to switch between Agent Metrics and Project Metrics.
- Create a Metric: The creation button will change based on your active tab (e.g., "Create Agent Metric" or "Create Project Metric").
- Convert an Agent Metric to Project-level: If you have an existing agent metric that you want to apply to all agents, click the three dots (...) next to the metric and select Move to Project Metric.
For more information on configuring these settings, please refer to our Basic and Advanced Metrics Guide.
4. What is the best way to perform A/B testing to compare two different conversational agents?
To perform effective A/B testing between two conversational agents (such as a production agent and a new version with a modified conversational flow) in Cekura, we recommend the following workflow:
-
Generate Test Cases (Evaluators): Start by generating a set of high-quality test cases. We recommend generating at least 10 good test cases that cover your agent's primary goals and edge cases. You can use Cekura's automated generation tools to create these based on your agent's context.
-
Establish a Baseline (Old Agent): Run these test cases against your existing 'old' agent first. This serves two purposes: it establishes a performance baseline and allows you to verify that the generated test cases are accurate and produce the expected results for your current flow.
-
Run Comparative Tests (New Agent): Execute the exact same set of test cases against your 'new' agent version.
-
Analyze and Compare: Once both runs are complete, navigate to the Results section in the Cekura dashboard. Select the two different test results you wish to evaluate, and then press the Compare button. This will provide a side-by-side comparison of how each agent performed against your defined metrics.
By running identical test cases against both agents, you can directly measure improvements or regressions in performance, accuracy, and flow adherence. For more information on setting up your agents for testing, refer to the Agent Setup Guide. To learn more about how to interpret the results, see our documentation on Basic and Advanced Metrics.
5. Is there a way to export or download call totals and topic breakdowns from the dashboard instead of manually copying and pasting?
Yes, Cekura provides several ways to access and export your call data, totals, and topic breakdowns without the need for manual copying and pasting.
1. Using the Cekura API
The most efficient way to export data for external reporting—such as emailing a daily breakdown of calls by topic—is through our API. The API allows you to programmatically fetch call data and metadata, which can then be formatted into your own reports or integrated into internal tools.
- Granular Metrics: The API often provides access to more detailed metrics than what is currently visible on the standard dashboard.
- Integration: You can leverage these results in your own web applications, allowing your team to focus specifically on calls that have been flagged for review.
- API Documentation: You can find the reference for fetching call data here: https://docs.cekura.ai/api-reference/observability/get-call.
2. Dashboard Reporting
While we are constantly improving the dashboard's export features, you can currently view detailed breakdowns by navigating to your agent's reporting page.
- Timeframe Selection: Ensure you use the timeframe selector (located at the top right of the dashboard) to filter for specific dates, such as "Today" or a custom range.
- Direct Links: If you encounter any issues with page loading or stalling, you can often access the specific data view by ensuring your URL includes the correct organization and agent IDs along with the
fromandtoISO timestamps.
3. Custom Support for Reporting
If you need a specific breakdown that is not currently documented in the public API reference, please reach out to the Cekura support team. We can provide specific endpoints tailored to your organization's reporting requirements to help you automate your customer communications.
6. Is there a way to update or add metadata to data after it has been posted to the observability endpoint?
Currently, it is not possible to modify or add metadata to a call record once it has been posted to the Cekura observability endpoint. Data sent to this endpoint is treated as immutable to maintain the integrity of the evaluation results and historical reporting.
Best Practices for Metadata
To ensure your analytics and metrics are as comprehensive as possible, we recommend the following:
- Collect All Data Upfront: Ensure that all relevant identifiers, such as CRM IDs, user segments, or session variables, are gathered before sending the payload to Cekura.
- Include Metadata in the Initial POST: All metadata should be included within the
metadataobject of your initial request to the observability API. - Verify Data Before Sending: Double-check that the metadata values are correct, as they will be used for filtering and calculating advanced metrics immediately after ingestion.
For more information on how to structure your metadata for optimal use in metrics, please refer to our documentation: https://docs.cekura.ai/documentation/key-concepts/metrics/basic-advanced-metrics#2-metadata
7. When making an inbound call to the system for testing, do I need to manually trigger the conversation?
The process for triggering a conversation depends on whether your agent is designed to receive calls (Inbound) or make calls (Outbound).
1. If your agent receives calls (Inbound Agent)
If you have an agent that receives calls, you do not need to manually dial. Simply provide the phone number associated with your agent in the Agent Settings. You can also override this number when you start a test run. Cekura will then initiate the call to your agent automatically.
2. If your agent makes calls (Outbound Agent)
If your agent is designed to make calls to people (which Cekura receives as inbound calls), the workflow is as follows:
1. Run Evaluators: Start the test run in Cekura first.
2. Get Phone Numbers: Cekura will provide you with a specific list of phone numbers to call.
3. Initiate Calls: You will need to make calls from your system to these provided numbers. You can leverage our APIs to automate this process.
3. Automatic Triggering via Integrations
If you use Retell, Vapi, or 11labs for your agent, Cekura can automatically trigger calls from your system to our numbers without manual intervention. To enable this:
1. Go to Agent Settings.
2. Select Voice Integration.
3. Choose your provider (e.g., Retell, Vapi, etc.).
4. Fill in the required details and enable the Outbound Auto Call flag.
For more detailed information on testing outbound agents, please refer to our documentation: Testing Outbound Calls.
8. How can I request a copy of your SOC 2 report?
To request a copy of our SOC 2 report, please reach out to us via support@cekura.ai or through your dedicated account manager. While the trust site does not explicitly mention the access process, you can find other relevant security and compliance information on our Trust Site here: https://tatva-labs-inc.trust.site/
9. How can I obtain a Business Associate Agreement (BAA) for HIPAA compliance?
Cekura supports HIPAA compliance for organizations that process Protected Health Information (PHI) during the testing and monitoring of AI voice and chat agents. To obtain a Business Associate Agreement (BAA), please follow these steps: 1. Reach out to your Cekura Account Manager or contact our support team. 2. Our legal department will provide a standard BAA for your review and signature. 3. Once executed, your organization can safely use Cekura to evaluate call recordings, transcripts, and metadata containing PHI. For example, healthcare providers can monitor patient-agent interactions to ensure accuracy and compliance without compromising data security.
10. How can I change the inbound number used to make calls to the main agent?
To change the inbound number used to make calls to your main agent, you need to update the configuration within the specific evaluator. Follow these steps:
- Edit Evaluator: Navigate to the specific evaluator you wish to modify within your scenario.
- Go to Configuration: Locate and click on the Configuration section.
- Update Phone Number: Find the phone number field and select your preferred number from the list of available phone numbers.
Note: If you want to use your own phone numbers (Bring Your Own Number), you can find instructions on how to integrate with Twilio here: https://docs.cekura.ai/documentation/key-concepts/phone-numbers/twilio
11. How can I integrate Cekura Cronjob success and failure notifications directly into a Datadog dashboard using webhooks?
To monitor the health of your Cekura Cronjobs within Datadog, you can leverage Cekura's webhook notification system. While Datadog does not currently provide a direct, generic webhook endpoint that can ingest Cekura's specific payload format without transformation, the integration can be achieved using an intermediary service.
Recommended Integration Workflow
-
Understand the Cekura Payload: Cekura sends detailed notifications regarding the success or failure of scheduled test runs. You can review the exact JSON structure of these notifications in the Cekura Webhook Documentation.
-
Create an Intermediary Webhook: Since Datadog requires data to be formatted for its specific Events or Logs API, you should set up a simple intermediary service (such as an AWS Lambda function, Google Cloud Function, or a lightweight server). This service will:
- Receive the POST request from Cekura.
- Extract relevant fields (e.g., job status, agent name, or failure reason).
- Transform and forward this data to the Datadog Events API. -
Configure Cekura Notifications:
- In the Cekura platform, navigate to the settings for your specific Cronjob.
- Provide the URL of your intermediary service in the webhook configuration field. -
Build the Datadog Dashboard: Once the events are flowing into Datadog, you can use the Datadog Dashboard builder to create visualizations. You can filter by the event source or custom tags you've sent from your intermediary script to track success/failure rates over time.
By using this approach, you ensure that your monitoring team receives real-time alerts and visual data regarding your AI agent testing schedules directly within your existing Datadog environment.
12. How do I choose a subscription package for load testing, and how can I perform load testing for inbound and outbound calls?
Cekura is built to handle high-scale performance testing, supporting north of 2000 concurrent calls.
Subscription Packages and Concurrency
Our subscription plans include specific concurrency limits. Higher concurrency requires higher committed resources from our infrastructure to ensure your tests run smoothly and accurately.
- Upgrading Concurrency: If the concurrency level you need to test exceeds your current package limits, please reach out to us at support@cekura.ai or contact your dedicated Account Manager to discuss an upgrade.
How to Perform Load Testing (Inbound and Outbound)
Whether you are testing inbound or outbound agents, the process involves simulating high traffic using your defined evaluators (test cases). Follow these steps to set up your load test:
- Select Evaluators: Pick one or more evaluators (test cases) that you want to run as part of the load test.
- Set the Frequency: To reach your desired concurrency level, you must distribute the load across your selected evaluators. Use the following formula:
- Frequency = Total Target Concurrency / Number of Evaluators Selected
- Execute the Test: Once configured, Cekura will simulate the calls—either by dialing out to your agent (outbound) or acting as the caller for your agent (inbound)—to stress-test the infrastructure and agent logic.
- Analyze Results: After the run, Cekura evaluates the performance against your defined metrics and expected outcomes to identify any infrastructure issues or false flagging.
For a comprehensive step-by-step walkthrough, please visit our Load Testing Documentation.
13. Can I define custom tools or scripts to perform setup and cleanup tasks before and after a Cekura test for end-to-end integration testing?
Yes, you can perform setup and cleanup tasks by integrating Cekura into your automated end-to-end testing workflows. This allows you to verify that your AI agent integrates correctly with your entire system architecture.
Recommended Workflow for Integration Testing
To achieve a full end-to-end test cycle, you can wrap the Cekura testing process within your own automation scripts or CI/CD pipeline:
- Setup (Prep Work): Run your custom scripts to prepare the environment. This might include database seeding, provisioning test users, or initializing specific system states.
- Initiate Cekura Test: Programmatically trigger a Cekura test run using our APIs. Cekura will use the agent context to generate test cases (evaluators), execute them, and evaluate the outcomes based on your defined metrics.
- Cleanup: Once the Cekura evaluation is complete, run your teardown scripts to reset the environment or clear test data.
By using this sequence, you can ensure that your agent performs as expected within the context of your broader system. For more information on programmatically interacting with the platform to trigger these runs, please refer to our API documentation at https://docs.cekura.ai/api-reference/.
14. Why are successful parallel outbound calls showing as timeouts in the dashboard?
Timeouts in the Cekura dashboard during outbound testing usually occur when the platform cannot successfully link an incoming call to an active test evaluator run. This is particularly common during parallel testing if the runs are not initialized correctly.
Common Causes for Timeouts
- Unmatched Evaluator Runs: Each individual call must correspond to a specific triggered evaluator run. For example, if you make three parallel calls but only trigger the "Run Evaluator" action once in the dashboard, Cekura will only record one run. The other calls will not be matched to a session, leading to timeouts for the expected test cases.
- The Outbound Timeout Window: By default, once you trigger an evaluator (via UI or API), you must initiate the call within 5 minutes. If the call is received after this window, the session expires and results in a timeout.
* Pro Tip: This window is configurable. You can adjust the duration by navigating to Settings -> Project -> General -> Outbound Timeout. - Caller ID Mismatch: Cekura identifies calls based on the phone number specified in your Agent Settings. If the incoming caller ID does not match the expected number for that specific test, the system cannot link the call to the evaluator.
- Mode Configuration: When running outbound tests, ensure you have selected the correct mode (
same_numbervsdifferent_numbers) for handling multiple concurrent calls.
Best Practices for Parallel Testing
To ensure all parallel calls are tracked accurately and to avoid manual errors:
- Use the API: Instead of manually copy-pasting numbers from the UI, use the Run Scenarios API. This allows you to programmatically trigger multiple runs and receive the specific destination numbers for each call in real-time.
- One-to-One Mapping: Ensure that for every call your system initiates, a unique request is sent to the Cekura "Run Evaluator" endpoint first. This creates the necessary placeholder in our system to receive and evaluate the call data.
If you continue to see timeouts despite these configurations, please verify that your SIP routing is correctly passing the caller ID to Cekura.
15. How can I configure a custom domain and logo for my reports, and what should I do if I encounter an error during the domain setup?
To white-label your reports and shareable links in Cekura, you can configure a custom domain and logo through the platform settings. Please note that custom domain configuration is an Enterprise-level feature and is not available on the Startup plan.
How to Configure Custom Branding
- Navigate to Settings in your Cekura dashboard.
- Go to the Domains section.
- Enter your domain details and upload your company logo.
Troubleshooting and Plan Restrictions
If you encounter an error while setting up your domain, please check the following:
- Plan Eligibility: Custom domains and logo configurations are restricted to Enterprise accounts and are no longer supported on the Startup plan. If you are on a Startup plan, the fields may appear in the UI, but the configuration will not be saved.
- Manual Workaround: If you need to provide a branded PDF report immediately and cannot update the logo via the dashboard, you can manually edit the Cekura logo from the generated PDF using a standard PDF editor.
For further assistance or to upgrade your plan, please visit the Cekura Documentation.
16. How can I perform end-to-end testing of a chatbot's LLM flow via SMS to ensure all conversation types are handled correctly?
Cekura provides native support for testing SMS-based chatbot flows, allowing you to validate the end-to-end logic of your AI agents. You can test these flows by either using Cekura's native SMS integration or by connecting your custom backend via an API or WebSocket.
How to Set Up SMS Testing
To begin testing your SMS agent, follow these steps in the Cekura platform:
- Configure the Agent: Navigate to Agent Settings > Chatbot.
- Select Provider: Choose SMS as the provider. If this option is not yet visible in your UI, please contact Cekura support to have it enabled for your account.
- Define Evaluators: Create new test cases (evaluators) or reuse existing ones that define the expected behavior and metrics for your conversation.
- Run the Test: In the testing interface, set the chat mode to SMS and click Run. Cekura will simulate the conversation and evaluate the performance based on your defined metrics.
Integration Options for Custom Backends
If you are using a custom backend or want to test the AI logic without relying on a carrier like Twilio, you have two primary options:
- API Endpoint: Expose an API endpoint that accepts a user's message and returns the agent's response. This allows Cekura to interact with your system as if it were sending and receiving SMS messages.
- WebSocket Integration: Cekura supports chat-based testing via WebSockets. You can use a sample script to convert your local logic into a WebSocket endpoint and route it through a tool like ngrok to provide Cekura with a public URL. You can find a reference implementation here: LLM WebSocket Server Example.
Key Benefits
- E2E Validation: Test the entire LLM logic from the initial user text to the final agent response.
- Automated Evaluators: Automatically generate test cases based on your agent's context to ensure comprehensive coverage of conversation types.
- Seamless Transition: Easily switch between voice and SMS testing if your system supports multi-channel dynamic transitions.
For more details on setting up your agent and defining metrics, please refer to the Cekura Agent Setup Guide.
17. How can I correlate incoming calls with the specific evaluation runs that triggered them in the Cekura API?
Correlating incoming calls to specific test runs is essential for organizing your evaluation data and ensuring each call is matched to the correct test case. Cekura provides two primary methods to achieve this: randomized phone numbers and custom integrations.
1. Randomized Phone Numbers (Immediate Correlation)
The simplest way to distinguish between concurrent calls is to enable the randomize phone number flag. By setting the mode to use different numbers in your request, each parallel test case will originate from a different caller ID. When calling the Run Scenarios endpoint, ensure your configuration is set to randomize the outbound numbers. This allows your system to match the incoming call's phone number to the specific run ID immediately upon receipt.
2. Custom Integration (Deterministic Correlation)
For a more robust and scalable solution, you can implement a custom integration. This allows you to link your internal system's unique identifiers (like a CallSid) with Cekura's evaluation runs.
- The Process: Once a call is complete on your end, you send Cekura the transcript and your local call ID (which Cekura labels as the
provider_call_id). - Automated Matching Logic: Cekura uses a deterministic matching algorithm to correlate your calls to the evaluation runs based on several factors: phone numbers, call durations, start times, and transcript fuzzy matching.
- Retrieval: You can then fetch specific runs using the List Runs endpoint and filtering by the
provider_call_idto find the corresponding Cekura run ID.
Note for Retell and Vapi Users
If you are using a supported provider like Retell or Vapi, you do not need to manually send the correlation data. If the integration is correctly configured in your Cekura dashboard, the provider_call_id will be automatically populated in your run results, allowing for seamless tracking without additional code.
18. What does the response consistency metric test for?
The Response Consistency metric is a pre-defined evaluation tool in Cekura designed to ensure your AI agent remains reliable and stable throughout a conversation. It primarily focuses on two key areas:
- Information Stability: This checks if the agent provides the same answer when asked the same question multiple times within a single session. For example, if a user asks for the price of a service three times, the agent should provide the exact same pricing information each time.
- Data Persistence and Accuracy: This verifies that the agent correctly remembers and utilizes information provided by the user. For instance, if a user states their phone number is "123-456-7890", the metric checks if the agent uses that same number when repeating it back or referencing it later in the interaction.
You can easily add this to your test cases or monitoring setup by selecting it from the list of pre-defined metrics. This is a critical check for building trust, as inconsistencies in pricing or data handling can lead to a poor user experience.
For more information on how to use metrics to evaluate your agent's performance, please visit our documentation: https://docs.cekura.ai/documentation/key-concepts/metrics/basic-advanced-metrics
19. How do I configure a Retell chat agent in Cekura and ensure that the chat tests appear on the dashboard?
To configure a Retell chat agent and ensure your test results are visible on the Cekura dashboard, follow these steps:
1. Prepare your Retell Agent
First, ensure you have a chat-compatible version of your agent. In your Retell dashboard, you can Copy as chat agent any of your existing voice agents to enable text-based interactions.
2. Connect Retell to Cekura
Next, link your agent to the Cekura platform:
- Navigate to Agent Settings on your Cekura dashboard.
- Go to the Chatbot Control tab.
- Select Retell as the provider.
- Enter the required connection details for your Retell agent.
3. Run and Monitor Tests
To execute your tests and ensure they are correctly logged to your dashboard:
- Navigate to your Evaluators.
- When initiating a test run, select the Run as Text option. This ensures the system treats the interaction as a chat session rather than a voice call.
Once the test is complete, the results will automatically populate your dashboard for analysis. For more specific details on this setup, please refer to the Retell Integration Guide.
20. What is the best way to evaluate a scenario where a caregiver or family member picks up instead of the patient, and should I adjust agent settings or generate evals for this?
To evaluate scenarios where a caregiver or family member answers the call instead of the patient, the best approach is to generate specific evaluators (test cases) using the "Extra Instructions" feature rather than tweaking your agent's core settings. This allows you to test the agent's robustness against different personas without altering its underlying logic.
Step-by-Step Instructions
- Navigate to Evaluator Generation: Within the Cekura platform, go to the section where you generate your test cases.
- Apply Extra Instructions: Use the "Extra Instructions" field to provide specific context for the AI persona.
- Use a Specific Prompt: Enter a prompt similar to the following:
"Generate a scenario where the caregiver or family member picks up instead of the patient. The caregiver should go through the flow on the patient's behalf."
- Run the Test: Execute the generated test cases to observe how your agent handles the interaction with a non-patient party.
Why this approach?
By generating evaluators with specific instructions, you can simulate real-world variability. This method tests if your agent can maintain the conversation flow and meet its objectives even when the primary contact is unavailable, which is a critical metric for healthcare-related AI agents.
21. Will incoming calls be restricted to a specific phone number, or can I still receive calls from multiple different numbers?
Cekura does not restrict incoming calls to a specific phone number; the platform is designed to be fully compatible with all numbers. Each evaluator (test case) has a phone number field, you can select from the pool of numbers available or bring your own phone numbers. You can continue to receive and test calls from multiple different sources, which is particularly useful for simulating diverse real-world scenarios for your AI voice agents. This flexibility ensures that your testing and monitoring workflows are not limited to a single originating caller ID. For example, when testing an agent, you can run test cases that originate from various numbers to ensure the agent's logic and metrics hold up across different caller contexts. Similarly, for monitoring, Cekura allows you to send call recordings or transcripts from any number to be evaluated.
22. Does Cekura support a webhook for receiving call observability metrics and evaluation results to store in an external database for analytics?
Yes, Cekura supports webhooks that allow you to receive real-time results of call observability metrics and evaluations. This feature is specifically designed to help teams integrate Cekura's insights into their own infrastructure, such as external databases or custom analytics dashboards.
How it Works
When Cekura finishes evaluating a call based on your defined metrics, it can automatically trigger a webhook. This sends a POST request containing the evaluation results and associated metadata to a destination URL of your choice.
Key Benefits
- Data Persistence: Store your evaluation results in your own database for long-term record-keeping.
- Custom Analytics: Use your preferred BI tools (like Tableau, Looker, or Grafana) to visualize Cekura's metrics alongside other business data.
- Automation: Trigger downstream workflows in your system based on specific evaluation outcomes.
Documentation
You can find the technical specifications, including the payload structure and setup instructions, in our documentation here: Webhook Format.
23. Which model does Cekura use to calculate the metric results?
Cekura primarily utilizes Gemini models to calculate metric results and perform evaluations across both its testing and monitoring workflows. These models are used to analyze agent interactions—whether they are generated during automated testing or captured from live call recordings and transcripts—against the specific metrics and expected outcomes defined in your setup. By leveraging Gemini's advanced reasoning capabilities, Cekura ensures high-quality and consistent evaluation of your AI voice and chat agents. For more information on how metrics are structured and used within the platform, you can refer to the documentation on Basic and Advanced Metrics.
24. How can I link an ElevenLabs account to view conversation IDs and tool call timestamps for evaluator test calls?
To link your ElevenLabs account and track granular conversation data such as Conversation IDs and tool call timestamps, follow these steps:
- Navigate to Agent Settings: In the Cekura dashboard, go to the settings for the specific agent you are testing.
- Select ElevenLabs: Under the provider or voice settings, select ElevenLabs.
- Provide API Access: Ensure you have provided the necessary API key permissions to allow Cekura to fetch conversation data.
Why link your ElevenLabs account?
- Conversation ID Tracking: Cekura will automatically match your evaluator test calls to the specific ElevenLabs Conversation ID.
- Tool Call Visibility: You will be able to see exactly when tool calls were performed during the interaction, making it easier to debug the agent's logic and response times.
25. How should I format transcript data for Cekura's observability API if my source only provides a single timestamp and not an end time?
Before manually modifying your data, first check if your current transcript format is natively supported by Cekura here: https://docs.cekura.ai/documentation/advanced/transcript-format. If your format is supported, you can send it as-is without manual transformation.
If your format is not supported and you only have a single timestamp available (as is common with some platforms like Trillet AI), you should follow these steps to format it for Cekura's Observability API:
Recommended Workaround
Map the single timestamp provided by your source to both the start and end time fields in the Cekura transcript object.
- Use the provided
timestampas thestart_time. - Set the
end_timeto be identical to thestart_time.
Important Note on Latency
By setting the start and end times to the same value, Cekura will not be able to provide accurate latency or duration metrics for those messages. While other evaluations (such as sentiment, accuracy, and goal completion) will work perfectly, the latency will be recorded as zero.
Example Transformation
Source Data:
{ "timestamp": "2025-09-02T10:36:59.505Z", "role": "agent", "content": "One moment while I send you the link" }
Cekura Format:
{ "role": "agent", "content": "One moment while I send you the link", "start_time": "2025-09-02T10:36:59.505Z", "end_time": "2025-09-02T10:36:59.505Z" }
26. How can I add Time to First Audio (TTFA) as an infrastructure metric for a Pipecat and Twilio setup?
In Cekura, you can track Time to First Audio (TTFA) and other infrastructure-related latencies by leveraging the platform's built-in metrics and transcript data. Cekura provides visibility into these metrics to help you monitor performance across your Pipecat and Twilio integrations.
Accessing Latency Metrics
You can access detailed latency data for your calls through Cekura's Python-based metric evaluation system. The latency_metrics object provides timing information for the conversation flow.
- TTFA Calculation: If your agent is configured to speak first (e.g., a greeting), the TTFA is represented by the
start_time(in seconds) of the first message from the main agent. - Reference: Cekura Latency Metrics Documentation
Handling Complex Scenarios with Transcript Data
In cases where the conversation flow is more complex—such as when a testing agent speaks first or you need to measure specific turn-around times—you should use the transcript_json directly.
- Granular Timing: The
transcript_jsoncontains a detailed breakdown of every message, including precisestartandendtimestamps for each turn. This allows you to programmatically calculate TTFA or any other custom latency metric based on the specific sequence of events in the call. - Reference: Cekura Transcript JSON Documentation
Implementation Workflow
To track these metrics for your Pipecat + Twilio agent:
1. Send Call Data: Ensure your call recordings, transcripts, and metadata are sent to Cekura via the observability APIs.
2. Define Metrics: Use the Python code metrics to extract the start_time of the first agent message from the latency data.
3. Monitor: View the resulting TTFA plots in your Cekura dashboard to identify trends or spikes in infrastructure performance.
27. How do I configure the Cekura bot to send results to a Slack channel if no results are appearing after adding it?
To configure the Cekura Slack integration and ensure results are appearing in your channel, follow these steps: 1. Go to Settings in the Cekura platform. 2. Ensure you are in the relevant project. 3. Navigate to Integrations and click Slack. 4. Follow the steps listed under 'How to set up Slack notifications' to connect your workspace. 5. Add the Cekura Bot: After the integration is set up, you must manually add the bot to your desired Slack channel. You can do this by typing /invite @Cekura in the channel or using the 'Add apps' option in Slack. 6. Enable Notifications: Make sure 'test result notifications' are enabled in your project settings. This configuration allows the bot to post evaluation results, which you can then use to reference specific runs and start discussion threads within Slack.
28. Is screen recording or video/microphone use disabled in the Cekura dashboard?
No, Cekura does not disable screen recording, video, or microphone usage on its dashboard. Users are free to use third-party recording tools (such as Loom) or system-level recording features while navigating the platform.
If you are experiencing issues with recording or media access, we recommend checking the following:
1. Browser Permissions: Ensure that your browser has granted the necessary permissions to your recording tool or the Cekura site.
2. System Privacy Settings: On macOS or Windows, verify that your recording application has permission to record the screen, microphone, or camera in the system settings.
3. Extension Conflicts: Sometimes other browser extensions can interfere with media capture. Try recording in an incognito window or disabling other extensions to troubleshoot.
If you continue to face difficulties, please contact our support team so we can investigate further.
29. Does the company provide a Data Processing Agreement (DPA) and what is the status of its GDPR compliance?
Cekura is fully GDPR compliant, ensuring that all data processing activities meet the rigorous standards set by the European Union for data protection and privacy. We provide a Data Processing Agreement (DPA) to our customers upon request to help satisfy legal and regulatory requirements. Please note that the DPA is provided upon request for a fee. To initiate this process or to receive further documentation regarding our compliance standards, please contact our support team.
30. How can I schedule runs of evaluators in the platform?
Yes, you can schedule automated runs of your evaluators using the Cronjobs feature in Cekura. This allows you to perform regular testing and monitoring of your AI agents without manual intervention. To set up a schedule, follow these steps: 1. Select Evaluators: Go to your evaluators list and check the boxes next to the ones you want to schedule. 2. Open Actions: Click the Actions button found in the top right corner of the dashboard. 3. Configure Cron Job: Select Cron jobs from the menu. 4. Set Schedule: Choose your desired frequency and timing for the runs. Once configured, Cekura will automatically execute these test cases based on your defined schedule and evaluate them against your metrics.
31. How do I configure an agent for text-based runs, and does it require a different configuration than voice-based runs?
Yes, text-based runs require a different configuration than voice-based runs. While voice-based simulations primarily use a phone number to initiate a call, text-based simulations rely on digital interfaces and integrations to communicate with your agent.
Configuration Options for Text-Based Runs
Cekura provides several flexible methods to connect to your agent for text-based testing:
- Direct Integrations: You can connect directly to popular AI voice and chat providers such as Retell, Vapi, and others.
- SMS Testing: Cekura supports testing agents via SMS for mobile-based chat scenarios.
- WebSocket Bridge: For custom or proprietary agents, you can configure a WebSocket URL in the agent settings. This URL acts as a bridge, allowing Cekura to send and receive text messages directly to your agent's logic.
How to Set Up Your Agent
- Navigate to Agent Settings: Open the specific agent you want to test in the Cekura dashboard.
- Select Connection Method: Choose the appropriate integration type (e.g., WebSocket, SMS, or a specific provider like Vapi).
- Input Connection Details: If using a WebSocket, provide the specific WebSocket URL. If using a direct integration, ensure your API keys or provider-specific identifiers are correctly mapped.
- Initiate the Run: When you go to
Configure Run, select the Text Based option. Cekura will then use your configured digital interface instead of attempting to dial a phone number.
Cross-Platform Testing
A significant advantage of the Cekura platform is that you can run the same evaluators (test cases) across both text and voice channels. This ensures that your agent's logic, knowledge, and guardrails remain consistent regardless of how the user interacts with it.
For more detailed information on setting up these connections, please refer to the Chat-Based Testing Documentation and the Agent Setup Guide.
32. Do you have a testing integration with Telnyx?
Currently, Cekura does not have a native API integration with Telnyx for advanced features like tool call testing. However, you can still fully test your Telnyx-based voice agents by using their assigned phone numbers.
How it Works
Cekura is designed to interact with AI agents directly via phone calls. If your Telnyx agent has a phone number, Cekura can dial it to execute test cases and evaluate the agent's performance.
Setup Steps
- Agent Context: Define the context and purpose of your agent within the Cekura platform.
- Phone Number: Provide the Telnyx phone number you wish to test.
- Evaluators: Generate or define test cases (evaluators) based on your expected outcomes.
- Run Tests: Cekura will call the number, simulate the conversation, and evaluate the results based on your defined metrics.
Important Considerations
- Tool Call Testing: Because this method uses a standard phone connection rather than a deep API integration, testing specific backend "tool calls" (verifying that the agent triggered a specific internal function) is not currently supported for Telnyx.
- Onboarding: If you are setting up an agent without a direct integration, you can follow the Onboarding Guide located in the bottom-left corner of the Cekura dashboard for a step-by-step walkthrough.
For more information on configuring your agent for testing, please refer to our Agent Setup Guide.
33. How can I set up a metric to track when the bot goes silent for more than a specific duration?
To accurately track instances where a bot goes silent, you should use the Infrastructure issues metric. It is important to avoid using an LLM-as-a-judge metric for this purpose, as LLMs evaluate transcript content and cannot reliably measure technical silence or specific time durations.
Configuration Steps
The Infrastructure issues metric is a statistical metric that is usually added as a default project metric in the Cekura platform. If you need to set it up or adjust it, follow these steps:
- Navigate to Metrics: Go to the metrics configuration section of your project.
- Add the Metric: If it is not already present, add the Infrastructure issues metric from the pre-defined metrics list.
- Edit the Duration: Click the Edit button next to the Infrastructure issues metric.
- Specify the Threshold: Enter the specific duration (e.g., 3 minutes) over which silence should be marked as a failure.
- Save: Save your changes to apply the metric to your monitoring or testing workflow.
Choosing the Right Metric
- Infrastructure issues: Use this metric when you specifically want to detect when the main agent is going silent for more than the configured time.
- Silence detection: If you care about detecting silence from either the main agent or the testing agent/user, use the Silence detection metric instead.
34. Does Cekura provide an API to measure latency and other observability metrics for calls?
Yes, Cekura provides comprehensive API support for observability, allowing you to measure latency and other critical conversational metrics for your AI voice and chat agents.
How Observability Works in Cekura
To monitor your calls, you can send conversation details—such as call recordings, transcripts, and any relevant metadata—to Cekura via our APIs. Once the data is ingested, Cekura evaluates the calls based on the metrics you have defined.
Key Capabilities:
- Latency Tracking: You can include latency data within the metadata sent to Cekura to monitor and analyze the responsiveness of your agents over time.
- Conversational Metrics: Beyond latency, you can define and track a wide range of metrics to evaluate the quality and effectiveness of the interaction.
- Data Retrieval: You can programmatically fetch call data and evaluation results using our API endpoints.
Relevant Resources:
- Fetching Call Data: To retrieve specific call details, you can use the Get Call API.
- Using Metadata: To understand how to leverage metadata (like latency) in your evaluations, see our Metadata Documentation.
35. What are the implementation requirements and costs for the 'voice tone + clarity' metric, and will an audio recording always be required for it to work?
The 'voice tone + clarity' metric is an audio-based evaluation tool designed to monitor the quality of AI voice agents. To use this metric, an audio recording is always required as the analysis is performed directly on the sound file. Implementation involves sending your call recordings along with any relevant metadata to Cekura for evaluation. The cost for using this metric is 0.2 credits per minute of audio recording processed. For more information on how to send call data via API, visit: https://docs.cekura.ai/api-reference/observability/get-call.
36. How can I access a list of available voice IDs and their configuration settings (such as gender, tone, and accent) to facilitate bias testing and personality overrides?
In Cekura, Personalities define the characteristics of the virtual caller, including their gender, accent, and tone. This feature is essential for bias testing, allowing you to observe if your AI agent behaves differently when interacting with various demographics.
How Personalities Work
By default, every test case (evaluator) you create has a specific personality attached to it. However, you do not need to manually edit each evaluator to test different voices. Instead, you can use the Personality Override feature during runtime to test the same scenario across multiple profiles.
Overriding Personalities at Runtime
To facilitate bias testing without modifying your existing evaluators, follow these steps:
1. Select Evaluators: In the Cekura dashboard, select the test cases you wish to run.
2. Initiate Run: Click the Run button.
3. Additional Configuration: In the run setup menu, navigate to the Additional Configuration section.
4. Select Personalities: Choose the specific personalities (e.g., American Female, British Male, Indian Female, etc.) you want to include in this test run.
Execution Logic
When you override personalities, Cekura performs one run per scenario for every personality selected. For example, if you select 10 evaluators and 3 different personalities, the system will automatically execute 30 total test calls and compile the results into a single report. This allows you to efficiently identify if gender or accent variations impact the agent's response logic or performance.
For more information on managing these profiles, please refer to the Personalities Documentation.
37. Can parameters from user profiles be included in the expected outcome of a manually defined scenario?
Yes, Cekura supports the use of Test Profile information within the Expected Outcome of your scenarios. This allows you to create dynamic assertions that validate whether an agent is correctly identifying or verifying specific user data.
How to Use Profile Parameters
You can access data from an attached test profile by using the placeholder syntax. For example, if you want to ensure the agent correctly validates a date of birth, you can include {{test_profile.dob}} in your expected outcome description.
Example Use Case: Verification Testing
If you are testing a verification workflow where a user might provide incorrect details, your expected outcome could be structured like this:
- Scenario: User provides a date of birth that does not match the profile.
- Expected Outcome: "The agent should compare the user's input against {{test_profile.dob}}. If the input is incorrect, the agent should ask for the DOB again. After two failed attempts, the agent must inform the user it cannot assist and direct them to call the facility directly."
Recommended Workflow
- Attach a Test Profile: Ensure the scenario is linked to a test profile containing the necessary metadata.
- Define Logic: Use placeholders to reference profile fields so the evaluator knows exactly what value the agent should be checking for.
- Automated Generation: You can also use the 'Generate Scenarios' feature with instructions like "Generate scenarios where the user provides fake verification information" while attaching a profile to automate this testing at scale.
For more technical details on accessing test profile data in your metrics and outcomes, please refer to the Basic and Advanced Metrics documentation.
38. Can I integrate and use my own custom ASR model within your simulation and observability environment for Saudi Arabic agents?
Yes, integrating custom ASR models for simulation and observability is supported. This feature is available exclusively on our Enterprise plans. Please reach out to our support team for more details on how to set this up.
39. Can I freeze or pause my subscription for a few weeks while my project is on hold?
Yes, if your project is on hold or you are between pilot phases, you can easily pause your subscription directly from the Cekura dashboard.
To pause your subscription:
1. Log in to your Cekura account.
2. Navigate to Billing.
3. Click on Manage Plan.
4. You will be redirected to our secure Stripe billing portal.
5. Select the option to pause your subscription.
You can return to this same section at any time to reactivate your plan and resume your AI agent testing and monitoring. This ensures that all your existing agent configurations, evaluators, and historical data remain intact for when you are ready to scale.
40. How can I bulk import scenarios and what file formats are supported?
Cekura supports the bulk import of scenarios (also referred to as test cases or evaluators) to help you scale your testing and monitoring efforts efficiently.
Supported File Formats
You can provide your scenarios in various formats. The most commonly used and recommended formats are:
- CSV: Ideal for simple, tabular scenario data.
- JSON: Best for structured data or complex test cases.
- Custom Formats: If your data is in a different format, the Cekura team can assist with internal conversion to ensure it is compatible with the platform.
How to Import Scenarios
API Import: You can programmatically import scenarios using Cekura's APIs. This is the most efficient method for teams looking to automate their testing pipeline. You can explore our API Reference for more technical details.
41. How do I add an extra member to my workspace?
Adding team members to your Cekura workspace allows for collaborative testing and monitoring of your AI agents. To invite a new member, follow these steps:
- Open Settings: Click on the Settings icon found in the top right corner of your dashboard.
- Go to Members: Select the Members tab from the navigation menu within the Settings area.
- Click Invite: Select the Invite button to open the invitation dialog.
- Send Invitation: Enter the email address of the team member you wish to add and send the invite.
Note: Please ensure you have the necessary administrative privileges, as workspace management typically requires admin access. If you are unable to see the 'Members' tab or the 'Invite' button, you may need to ask an existing administrator to update your permissions or send the invitation for you.
42. How can I interface a custom-hosted voice agent using the OpenAI Realtime API with Cekura to run evaluations?
Cekura provides multiple ways to interface with your voice agent for testing and evaluation, depending on your hosting infrastructure and communication protocol. To run evaluations against a custom-hosted agent (such as one using the OpenAI Realtime API), you can use one of the following connection methods:
1. WebRTC (LiveKit & Pipecat)
If your agent is hosted using standard WebRTC protocols, Cekura offers native support for:
- LiveKit: This is a common choice for OpenAI Realtime API implementations. You can follow our LiveKit Integration Guide to connect your hosted agent to Cekura.
- Pipecat: Cekura also supports the Pipecat protocol for seamless audio streaming and evaluation.
2. Telephony and SIP
For agents that are accessible via traditional voice channels, Cekura supports:
- Telephony: Testing via standard phone numbers.
- SIP: Connecting directly to your VoIP infrastructure via Session Initiation Protocol.
3. Custom Implementations
If you use a proprietary WebRTC stack or a custom audio streaming method that does not fall into the categories above, Cekura can build a custom connector to support your specific architecture. This ensures that Cekura can correctly feed audio to and receive audio from your agent for high-fidelity testing.
The Testing Workflow
Once the connection is established, you can follow the standard Cekura workflow to evaluate your agent:
1. Define Context: Provide the agent's purpose and connection type to Cekura.
2. Generate Evaluators: Cekura will automatically generate test cases (evaluators) or allow you to define your own metrics.
3. Run Tests: Cekura interacts with your agent through the chosen connection type (WebRTC, SIP, etc.).
4. Evaluate Results: The platform analyzes the conversation based on defined metrics and expected outcomes to provide a performance report.
43. How can I prevent a testing agent or evaluator from automatically ending a call when it is instructed not to disconnect?
If your testing agent (evaluator) is ending calls prematurely despite having specific instructions in the scenario prompt to wait for the main agent to disconnect, you may need to disable the End Call tool in the evaluator's configuration. This ensures the evaluator does not have the technical capability to hang up, regardless of the conversation flow.
Steps to Disable End Call
- Navigate to Evaluators: Log in to your Cekura dashboard and go to the Evaluators section.
- Edit Evaluator: Find the evaluator you are using for the test and click on Edit.
- Access Configuration: Within the evaluator settings, navigate to the Configuration tab.
- Modify Tool Calls: Look for the Tool Calls section.
- Disable end_call: Locate the
end_calltool and switch it to Disabled.
By disabling this tool, the testing agent will be forced to stay on the line until the main agent terminates the session.
Note - It is not recommended to disable this tool. Instead provide better instructions on when to end the call.
44. How do I set up outbound tests?
Setting up outbound tests in Cekura allows you to evaluate how your AI voice or chat agents perform when initiating contact, such as for appointment reminders or follow-up calls. To set up these tests, follow these steps: 1. Define Agent Context: Provide the background, purpose, and knowledge base of your agent to help Cekura understand the expected behavior. 2. Generate Evaluators: Create test cases (evaluators) manually or use Cekura's auto-generation feature to define success criteria and expected outcomes for the interaction. 3. Configure Outbound Parameters: Set up the specific triggers and destination details for the outbound reach-out. 4. Execute and Evaluate: Run the test cases and review the results based on defined metrics to ensure your agent is performing as expected. For a comprehensive step-by-step guide on configuring these tests, please refer to our official documentation: https://docs.cekura.ai/documentation/guides/testing-agents/outbound-evaluators
45. How can I limit the number of concurrent evaluator runs to manage parallel call capacity?
To manage parallel call capacity and avoid hitting concurrency limits from providers (such as VAPI), Cekura allows you to set a maximum limit for concurrent evaluator runs at the organization level. This ensures that when you trigger a batch of test cases, the system respects your infrastructure's constraints.
How to configure the Parallel Call Limit:
- Log in to your Cekura dashboard.
- Navigate to Settings in the sidebar.
- Under the Organisation section, select General.
- Locate the field labeled Parallel Call Limit.
- Enter the maximum number of concurrent calls or evaluators you want to allow (e.g., if your VAPI limit is 10, you might set this to 10).
- Save your changes.
Once this is set, Cekura will automatically queue and throttle your evaluator runs to ensure they do not exceed the specified limit, preventing errors related to rate limiting or concurrency issues during your testing workflow.
46. How can I test LiveKit voice agents that do not have phone numbers attached using Cekura?
Yes, Cekura supports testing LiveKit voice agents directly via WebRTC, allowing you to run automated tests even if your agent does not have a phone number or PSTN endpoint attached. This is particularly useful for testing agents used in web or mobile applications.
How to Set Up LiveKit WebRTC Testing
To test your LiveKit agent, you will need to integrate your LiveKit environment with Cekura. Follow these steps:
- Access Integration Settings: Navigate to the LiveKit integration guide in the Cekura documentation to understand the required parameters.
- Provide Credentials: You will typically need your LiveKit URL, API Key, and API Secret to allow Cekura to initiate WebRTC sessions.
- Define Agent Context: Just like phone-based testing, provide the context and instructions for your agent so Cekura knows how to interact with it.
- Generate Evaluators: Cekura will automatically generate test cases (evaluators) based on your agent's context, or you can manually define specific scenarios you want to test.
- Run and Evaluate: Execute the test cases. Cekura will connect to your LiveKit agent via WebRTC, conduct the conversation, and evaluate the performance based on your defined metrics.
Billing and Credits
Testing LiveKit agents via WebRTC utilizes credits in the same way as standard phone call testing. There is no difference in the billing rate between PSTN and WebRTC-based tests.
For a detailed step-by-step walkthrough on setting this up, please refer to our official guide: LiveKit Integration Guide.
47. How can I configure daily reports to be sent to a specific email address instead of all project members?
To route daily reports to a specific email address (such as a shared inbox or a Google Group) rather than all project members, you need to ensure that the target email is registered as a member of your project. Follow these steps to configure your reporting preferences:
1. Invite the Target Email Address
Currently, Cekura sends reports to active members within the organization. You must invite the specific email address you want to receive reports to your Cekura account.
2. Create an Account for the Email
The invited email must have an active Cekura account to receive notifications.
- If using a shared inbox or Google Group: Since these often cannot use SSO (Single Sign-On), you should create the account using the Email/Password signup method.
- Ensure the invitation is accepted and the user appears in your member list.
3. Disable Notifications for Other Members
Once the designated reporting email is added, you can restrict who receives the daily updates:
- Navigate to Settings > Project > Members.
- Locate the members who should no longer receive the reports.
- Toggle off the email notification setting for those specific users.
Alternative: Email Forwarding
If you do not wish to add a new member to your Cekura project, you can set up an auto-forwarding rule within your own email client. You can configure your inbox to automatically forward any emails received from Cekura to your desired reporting address (e.g., pg-prod-reporting@prodigaltech.com).
Note: If you find that certain members (such as Admins) are still receiving emails after disabling them in settings, please reach out to support, as this may be a known synchronization issue regarding administrative permissions.
48. How can I add the observability URL to an n8n flow instead of configuring it directly in Retell?
If you already have a post-call analysis workflow in n8n using Retell's webhook slot, you can integrate Cekura observability by forwarding the call payload from your n8n flow to Cekura's endpoint. This allows you to maintain your existing logic while still benefiting from Cekura's monitoring and evaluation capabilities.
Step-by-Step Integration via n8n:
- Receive the Webhook: Ensure your n8n workflow starts with a Webhook node that receives the
call_analyzedorcall_endedevent from Retell. - Process Data (Optional): Perform any custom post-call analysis or data transformation within your n8n flow as you normally would.
- Add an HTTP Request Node: At the end of your flow (or at the desired point), add an HTTP Request node to forward the data to Cekura.
- Configure the Node:
- Method: POST
- URL: Use your project-specific Cekura observability URL.
- Body Parameters: Send the JSON payload received from Retell. Cekura needs the transcript, recording URL, and any relevant metadata to perform evaluations. - Authentication: Include your Cekura API key in the headers if required by your specific project configuration.
Helpful Resources:
- Documentation: For specific details on how to structure the forwarded response, please refer to the Cekura Retell Observability Guide.
- Video Tutorial: You can also view a community-created walkthrough of this setup here: Cekura Integration Video.
49. Where else do I need to update the billing email address if invoices are still being sent to the previous email after an update?
If you have updated your account email but invoices are still being sent to a previous address, you likely need to update the contact information in the Billing Portal. Cekura separates account profile settings from billing and subscription management to allow for different administrative and financial contacts. How to update your billing email: 1. Log in to the Cekura platform and navigate to Settings. 2. Click on the Billing or Subscription tab. 3. Select Manage Subscription or Update Billing Info to access the billing portal. 4. In the portal, find the Billing Information section and update the email address to your preferred recipient. 5. Save your changes. This ensures all future invoices and receipts are sent to the updated address. If the issue persists, please contact support for manual verification.
50. How can I replay evaluation webhooks for calls that were already evaluated prior to the webhook integration?
Currently, Cekura does not support a direct 'replay' feature for webhooks on evaluations that occurred before the webhook integration was active. Webhooks are designed to send real-time notifications for new evaluation events as they happen. To populate your system with data from calls that have already been evaluated, you should use the Cekura API to fetch the historical data. You can retrieve details for past calls using the Cekura Observability API at https://docs.cekura.ai/api-reference/observability/get-call. Since the data structure returned by the API is consistent with the information sent in a webhook payload, you can pass the API response through your existing webhook handler or ingestion logic to ensure your records are up to date. For all calls evaluated after your integration, the webhooks will trigger automatically, ensuring your system stays synchronized in real-time.
51. Is there a way to setup multiple phone numbers for a single agent?
Yes, Cekura allows you to use multiple phone numbers for a single agent by using the phone number override feature during test execution. This is particularly useful for CI/CD pipelines where you might have different phone numbers assigned to different environments (e.g., dev, staging, and production) but want to test the same agent logic.
How it Works
Instead of creating separate agents for every phone number, you can maintain a single agent configuration and specify which phone number to dial at the time you trigger a test run.
Benefits
- Simplified Management: You don't need to duplicate agent settings or evaluators for different environments.
- Version Control: Ensure that the version of the agent currently on your
devbranch is being tested against thedevphone number before it is merged tomain. - Dynamic Testing: Easily switch between different entry points for your AI voice agent without manual configuration changes in the Cekura dashboard.
52. How can I set up a custom metric with a numerical rating scale instead of a boolean pass/fail to capture more nuance in transcription accuracy evaluations?
While Cekura defaults to boolean (Pass/Fail) metrics because LLMs are generally more accurate with binary tasks, you can create custom Rating metrics to capture more nuance in your evaluations. Cekura supports a 0-5 rating scale for this purpose.
Steps to Create a Rating Metric
- Navigate to Metrics: Go to the 'Metrics' section in your Cekura dashboard.
- Create New Metric: Click the Create Metric button.
- Select Type: Choose Rating as the metric type. Please note that Cekura uses a 0-5 scale rather than 0-10.
- Describe the Metric: Provide a brief, clear description of what the metric is evaluating (e.g., 'Assess the accuracy of specific data points like dates and numbers mentioned in the transcript').
- Optimize: Click the Improve button. This allows the platform to refine the prompt logic to ensure the LLM understands the nuances of your request.
- Review and Assign: Review the metric configuration and add it to your existing test scenarios.
Best Practices for Numerical Scoring
- Define Score Meanings: To get the most accurate results, clearly define what each numerical value represents in your description. For example, specify what constitutes a '3' versus a '5'.
- Generic Use Cases: Custom rating metrics are most effective when the criteria are generic enough to be applicable across all or most of your calls.
For more detailed information on configuring metrics, you can refer to the Cekura Metrics Documentation.
53. How can I access an overall report of the results?
1. Accessing Overall Reports
You can view the summary and performance of your test runs in two main areas:
- Result Page: This page provides a detailed breakdown of a specific run, including individual test cases and specific action items to improve your agent.
- Runs Overview Page: This provides a high-level view of all your historical runs, allowing you to track progress over time.
54. How can I accurately track and report call quality metrics such as ghost inputs, transcript errors, and call termination reasons using the Cekura observability API?
To accurately track and report advanced call quality metrics in Cekura, you should leverage the observability API to send detailed call data, including specific metadata and termination reasons. This allows Cekura to evaluate the performance of your AI agent against expected outcomes.
1. Tracking Call Termination (User vs. Bot Hangups)
To improve the accuracy of the Appropriate Call Termination metric, you should explicitly send the reason the call ended. This helps Cekura distinguish between a user hanging up, the agent ending the call, or technical failures.
- Field:
call_ended_reason - Recommended Values:
user ended,agent ended,connection lost, etc. - Impact: Providing this field significantly improves the quality of termination metrics and helps identify if a bot generated a prompt that the user never heard due to a hangup.
- Reference: Cekura API - Send Calls
2. Handling Transcript Errors and ASR Inaccuracies
If you notice duplicate lines or inaccuracies in transcripts (often caused by ASR providers like Deepgram), ensure you are sending transcripts in a supported format. You can then leverage our "Transcription Accuracy" metric to detect inefficiencies in your transcript.
- Best Practice: Use the
transcript_jsonfield to send structured data rather than raw text. - Documentation: Transcript Format Guide
4. Using Metadata vs. Dynamic Variables
Understanding how to pass extra data is crucial for fine-tuning evaluations:
* Metadata: Use this for call-specific information (e.g., due_amount). During evaluation, Cekura can check if the bot correctly mentioned the specific due_amount passed in the metadata.
* Dynamic Variables: These are used to replace variables within your Agent Description. This is specifically used for the Instruction Follow metric to ensure the agent is adhering to its persona based on changing parameters.
55. Why are transcripts with tool calls not being fetched from Retell, and could this be due to HIPAA compliance being enabled?
If you notice that transcripts or tool calls are missing from your Retell-integrated calls in Cekura, it is highly likely due to HIPAA compliance settings being enabled on your Retell account.
Why this happens
When HIPAA compliance is active on a platform like Retell, it often restricts the storage of sensitive data or prevents transcripts and tool call logs from being accessed via external APIs to ensure data privacy. Since Cekura requires these transcripts to perform evaluations and monitoring, these restrictions prevent the platform from fetching the necessary data for your dashboard.
Troubleshooting Steps
- Verify Retell Settings: Log in to your Retell dashboard and check if HIPAA compliance is currently enabled.
- Check Data Permissions: If HIPAA compliance is necessary for your use case, ensure that your configuration allows for the secure transmission of transcripts to authorized third-party platforms like Cekura.
- Tool Call Metadata: Because tool calls are typically embedded within the transcript or call metadata, any restriction on the transcript will also result in missing tool call information.
For more details on how Cekura processes call data for monitoring, you can visit our Observability Overview. If you are manually fetching call data via our system, please refer to the Get Call API documentation.