Skip to main content

What are Evaluations?

Evaluations are automated tests that assess your agent’s performance after each call. They help you measure success rates, identify issues, and maintain quality standards across all conversations.
Evaluations provide quantitative metrics to track your agent’s performance over time.

How Evaluations Work

Each evaluation is a yes/no question that the AI analyzes based on the conversation transcript. After every call, Meetzy runs all configured evaluations and stores the results.

Example Evaluations

EvaluationQuestion
Appointment ScheduledDid the customer schedule an appointment?
Customer SatisfiedDid the customer express satisfaction with the service?
Issue ResolvedWas the customer’s issue successfully resolved?
Information CollectedDid the agent collect all required information?
Objection HandledDid the agent successfully address customer objections?

AI Copilot Suggestions

The AI Copilot can automatically suggest performance metrics and evaluation criteria based on your agent’s purpose and conversation patterns.

Automatic Evaluation Recommendations

The AI Copilot analyzes your agent configuration and suggests relevant evaluations:
  1. Context-Aware Suggestions: Based on your agent’s role (sales, support, appointments)
  2. Conversation Analysis: Recommendations from analyzing test conversations
  3. Industry Best Practices: Standard metrics for your use case
  4. Performance Optimization: Evaluations that identify improvement opportunities

Using Copilot Suggestions

1

Open AI Copilot

Click the sparkles icon in the right panel to access AI Copilot.
2

Request Evaluation Suggestions

Type something like “Suggest evaluations for my agent” or use the quick actions.
3

Review Recommendations

The Copilot will analyze your agent and suggest specific evaluation criteria.
4

Accept or Customize

Accept suggestions directly or modify them to fit your specific needs.

Copilot Suggestion Types

Evaluations focused on primary objectives:
  • “Did the agent achieve the call objective?”
  • “Was the primary goal accomplished?”
  • “Did the conversation meet success criteria?”

Applying Copilot Suggestions

When the AI Copilot suggests evaluations, they appear in a purple suggestion banner:
Review Copilot suggestions carefully before accepting them. While AI recommendations are based on best practices, they should be customized for your specific use case.
1

Review Suggestions

Examine the suggested evaluation name and question for relevance.
2

Customize as Needed

Modify the evaluation question to match your specific requirements.
3

Accept or Reject

Use the “Accept All” button to add all suggestions, or reject to dismiss them.
4

Fine-tune

Edit accepted evaluations in the main interface to perfect them.

Creating Evaluations

1

Open Evaluations Section

Navigate to the Evaluations section in the left panel (✅ icon).
2

Add Evaluation

Click “Add Evaluation” to create a new test.
3

Enter Name

Give your evaluation a descriptive name.
4

Write Question

Create a clear yes/no question that can be answered from the transcript.
5

Save

Save the evaluation to add it to your agent.

Evaluation Properties

PropertyDescriptionExample
NameShort identifier”Appointment Scheduled”
QuestionYes/no prompt for AI”Did the customer schedule an appointment during this call?”

Writing Effective Questions

The clearer your question, the more accurate the evaluation results will be.

Good Questions

  • ✅ “Did the customer agree to schedule a follow-up call?”
  • ✅ “Did the agent verify the customer’s contact information?”
  • ✅ “Was the customer’s primary concern addressed during the call?”
  • ✅ “Did the agent mention the current promotion?”

Poor Questions

  • ❌ “Was the call good?” (too vague)
  • ❌ “How satisfied was the customer?” (not yes/no)
  • ❌ “Rate the agent’s performance” (not answerable from transcript)

Copilot-Enhanced Question Writing

The AI Copilot can help improve your evaluation questions:
  1. Clarity Enhancement: Makes vague questions more specific
  2. Binary Conversion: Converts complex questions into yes/no format
  3. Context Addition: Adds relevant context for better evaluation accuracy
  4. Best Practice Alignment: Ensures questions follow evaluation best practices

Using Templates

Meetzy provides pre-built evaluation templates for common use cases:
  • “Did the customer express interest in the product?”
  • “Was pricing discussed during the call?”
  • “Did the agent attempt to close the sale?”
  • “Were all customer objections addressed?”

Applying Templates

1

Click Templates

Click the “Templates” button in the Evaluations section.
2

Browse Categories

Select a category relevant to your use case.
3

Select Evaluations

Choose which evaluations to add.
4

Customize

Modify the questions to match your specific needs.

Managing Evaluations

Editing Evaluations

  1. Click the edit icon (✏️) next to any evaluation
  2. Modify the name or question
  3. Save changes

Deleting Evaluations

  1. Click the delete icon (🗑️) next to the evaluation
  2. Confirm deletion
Deleting an evaluation removes it from future calls. Historical evaluation data is preserved.

Reordering Evaluations

Drag and drop evaluations to change their display order.

Viewing Results

Evaluation results are available in multiple places:
LocationWhat You See
Call DetailsIndividual evaluation results for each call
Analytics DashboardAggregated success rates over time
Webhook OutputRaw evaluation data sent to your systems
API ResponseProgrammatic access to results

Best Practices

Start with Copilot Suggestions

Begin by asking the AI Copilot to suggest evaluations. This provides a solid foundation based on your agent’s configuration and industry best practices.

Evaluation Quality Guidelines

  1. Be Specific: The more specific your question, the more reliable the evaluation
  2. Avoid Overlap: Don’t create multiple evaluations that measure the same thing
  3. Start Simple: Begin with 3-5 core evaluations that measure your most important KPIs
  4. Review Regularly: Periodically review evaluation results to calibrate questions
  5. Use for Training: Low-scoring evaluations highlight areas for agent improvement

Copilot Integration Tips

  • Iterate with AI: Use the Copilot to refine and improve existing evaluations
  • Context Matters: Provide context about your business when asking for suggestions
  • Test Suggestions: Use the Playground to validate Copilot-suggested evaluations
  • Combine Sources: Mix Copilot suggestions with templates and custom evaluations

Common Evaluation Patterns

Goal Achievement

Name: Primary Goal Met
Question: Did the call achieve its primary objective (sale, appointment, resolution, etc.)?

Compliance

Name: Disclosure Made
Question: Did the agent provide the required legal disclosure at the beginning of the call?

Quality

Name: Professional Tone
Question: Did the agent maintain a professional and courteous tone throughout the call?

Data Collection

Name: Contact Info Verified
Question: Did the agent verify or collect the customer's email and phone number?

Next Steps