Overview
Testing & Quality Assurance¶
Comprehensive testing, evaluation, and quality assurance framework for ART Voice Agent Accelerator.
Testing Strategy Overview¶
The ART Voice Agent Accelerator provides a multi-layered testing strategy to ensure quality at every level:
Quick Navigation¶
-
Unit & Integration Tests
Fast, isolated tests for core components and event handling
-
Agent Evaluation
Measure LLM quality: tool precision, groundedness, latency
-
Load Testing
WebSocket performance testing with Locust
Test Categories at a Glance¶
| Category | Purpose | Tools | Run Time |
|---|---|---|---|
| Unit Tests | Component isolation | pytest | Seconds |
| Integration Tests | Cross-component flows | pytest | Seconds |
| Agent Evaluation | LLM quality metrics | Evaluation framework | Minutes |
| Load Tests | Performance at scale | Locust | Minutes-Hours |
Evaluation Framework Highlights¶
The evaluation framework measures agent quality across multiple dimensions:
| Metric | Description |
|---|---|
| Precision | Fraction of tool calls that were correct |
| Recall | Fraction of expected tools that were called |
| Efficiency | Avoidance of redundant tool calls |
| Metric | Description |
|---|---|
| Groundedness | Response accuracy against evidence |
| Verbosity | Token usage and conciseness |
| Handoff Accuracy | Correct agent routing |
| Metric | Description |
|---|---|
| E2E Latency | End-to-end response time |
| TTFT | Time to first token |
| Cost | Token usage and estimated USD |
Running Tests¶
Quick Commands¶
# Unit tests
pytest tests/ -v
# Evaluation scenarios
pytest tests/evaluation/test_scenarios.py -v
# Load tests
make run_load_test_acs_media
Execution Methods¶
Run evaluations directly on your machine:
Run evaluations in GitHub Actions:
Documentation Structure¶
docs/testing/
├── index.md # This overview page
└── evaluation.md # Evaluation framework guide
docs/operations/
├── testing.md # Unit & integration tests
└── load-testing.md # Locust load testing
Getting Started Paths¶
I want to run unit tests
: Start with Testing Framework → pytest tests/ -v
I want to evaluate agent quality : Read Evaluation Framework
I want to load test my deployment
: Follow Load Testing Guide → make run_load_test_acs_media
I want to create custom scenarios : Check scenario examples in Evaluation Framework