Scaling Voice Assistant Testing for the Software-Defined Vehicle (SDV): How AI Can Enhance Traditional Test Methods
- アダム・ジェファーソン
- 3日前
- 読了目安: 3分
更新4時間前

For years, in-car voice assistants were largely transactional. Drivers issued rigid commands (“navigate home”) and the system executed a predefined action. Recent technological advances have fundamentally changed this interaction model. Conversational dialogue, contextual follow-up questions, and multi-modal interactions are becoming increasingly common. Voice assistants are evolving from command interfaces into conversational copilots capable of managing complex tasks across multiple vehicle domains.
SBD has been evaluating voice solutions for more than 14 years, assessing systems across multiple dimensions of the user experience. Historically, much of this testing process relied on manual activities and worked effectively when voice assistants were largely deterministic.
However, LLM-based systems introduce significantly greater variability and conversational depth, making traditional testing approaches increasingly difficult to scale. Simply expanding test coverage quickly becomes impractical due to the time required for documentation, transcription, and analysis.

To address this challenge, SBD has evolved its global voice testing framework to incorporate AI-assisted analysis while retaining critical human evaluation. The central question guiding this evolution was simple: Can testing become more efficient while simultaneously increasing the depth of insights?
The updated methodology separates the testing process into three layers.
Stage | Responsible | Role |
Voice interaction | Human tester | Generates natural voice prompts |
Data capture & analysis | AI agent | Transcribes audio, populates data sheets, performs initial scoring and analysis according to defined rules |
Validation | Human expert | Confirms results and provides contextual interpretation |
This approach preserves the human element required to effectively evaluate HMI quality, while automating the repetitive tasks that traditionally slow down testing. The AI agent processes human test recordings using a structured prompt that instructs it to:
Generate a full transcript of the interaction
Populate structured evaluation spreadsheets
Produce initial scoring against predefined criteria
Provide diagnostic insights and recommendations

Does It Work?
Initial trials demonstrate that the hybrid testing model reduces processing time while increasing the depth of analysis. Key advantages include:
Significant testing scale improvements
Automation reduces the time required to process results, enabling larger test datasets.
More consistent scoring
Structured prompts and automated transcription reduce variability in documentation.
Deeper insights
Automated analysis highlights failure patterns and improvement opportunities that may otherwise be missed.
Expert-verified outputs
Human validation ensures that complex conversational behaviour is interpreted correctly.
Testing at (Almost) the Speed of Software
As vehicles become increasingly software-defined, the pace of feature development continues to accelerate. This raises a natural question: can AI fully replace human testing?
In practice, fully automated testing risks missing the subtleties of human-machine interaction. Conversational interfaces rely heavily on context, tone, and intent interpretation – factors that remain difficult to evaluate without human judgement.
SBD’s hybrid approach – combining human interaction, AI-driven analysis, and expert validation – aims to bridge the gap between testing rigour and development speed.
Initial trials show that this methodology produces more detailed insights while significantly improving testing efficiency. The framework will continue to evolve throughout 2026 as voice assistants become an increasingly critical component of the in-vehicle user experience.

Why This Matters for OEMs
Voice is on the brink of becoming a core differentiator across brands, impacting perceived product quality and cockpit usability. However, as conversational systems grow more complex and development cycles accelerate, traditional testing methods struggle to keep pace. This creates a risk of inconsistent user experiences and missed edge cases.
OEMs that scale their voice testing capability can:
Deliver more consistent, high-quality user experiences
Identify issues earlier in development
Support faster, lower-risk software releases
SBD’s hybrid testing approach enables OEMs to match testing speed with software development while maintaining the depth of insight required for effective HMI evaluation.
"As the industry shifts towards in-car AI integration, test engineers must also leverage this new technology. Test complexity is growing and it is ever more critical to strike an effective balance between AI efficiency and human understanding.
In this insight, SBD looks at how voice recognition test procedures can be optimised with AI."




