Skip to main content

Documentation Index

Fetch the complete documentation index at: https://crewai-lorenze-imp-docs-improvements.mintlify.app/llms.txt

Use this file to discover all available pages before exploring further.

Overview

Testing is a crucial part of the development process, and it is essential to ensure that your crew is performing as expected. With crewAI, you can easily test your crew and evaluate its performance using the built-in testing capabilities.

When to Use Testing

  • Before promoting a crew to production.
  • After changing prompts, tools, or model configurations.
  • When benchmarking quality/cost/latency tradeoffs.

When Not to Rely on Testing Alone

  • For safety-critical deployments without human review gates.
  • When test datasets are too small or unrepresentative.

Using the Testing Feature

Use the CLI command crewai test to run repeated crew executions and compare outputs across iterations. The parameters are n_iterations and model, which are optional and default to 2 and gpt-4o-mini.
crewai test
If you want to run more iterations or use a different model, you can specify the parameters like this:
crewai test --n_iterations 5 --model gpt-4o
or using the short forms:
crewai test -n 5 -m gpt-4o
When you run the crewai test command, the crew will be executed for the specified number of iterations, and the performance metrics will be displayed at the end of the run. A table of scores at the end will show the performance of the crew in terms of the following metrics:
Tasks/Crew/AgentsRun 1Run 2Avg. TotalAgentsAdditional Info
Task 19.09.59.2Professional Insights
Researcher
Task 29.010.09.5Company Profile Investigator
Task 39.09.09.0Automation Insights
Specialist
Task 49.09.09.0Final Report CompilerAutomation Insights Specialist
Crew9.009.389.2
Execution Time (s)126145135
The example above shows the test results for two runs of the crew with two tasks, with the average total score for each task and the crew as a whole.

Common Failure Modes

Scores fluctuate too much between runs

  • Cause: high sampling randomness or unstable prompts.
  • Fix: lower temperature and tighten output constraints.

Good test scores but poor production quality

  • Cause: test prompts do not match real workload.
  • Fix: build a representative test set from real production inputs.