CoMAI: A Collaborative Multi-Agent Framework for Robust and Equitable Interview Evaluation
About
Ensuring robust and fair interview assessment remains a key challenge in AI-driven evaluation. This paper presents CoMAI, a general-purpose multi-agent interview framework designed for diverse assessment scenarios. In contrast to monolithic single-agent systems based on large language models (LLMs), CoMAI employs a modular task-decomposition architecture coordinated through a centralized finite-state machine. The system comprises four agents specialized in question generation, security, scoring, and summarization. These agents work collaboratively to provide multi-layered security defenses against prompt injection, support multidimensional evaluation with adaptive difficulty adjustment, and enable rubric-based structured scoring that reduces subjective bias. Experimental results demonstrate that CoMAI achieved 90.47% accuracy, 83.33% recall, and 84.41% candidate satisfaction. These results highlight CoMAI as a robust, fair, and interpretable paradigm for AI-driven interview assessment.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Interview Scoring and Admission Analysis | Interview Score Distribution (averaged across 55 participants) | Mean Score62.92 | 8 | |
| Interview Assessment | University Admission Interview Dataset 55 candidates (test) | Recall95.45 | 7 | |
| Adversarial Prompt Injection Defense | 500 adversarial samples | Defense Success Rate100 | 4 | |
| Interview Assessment | Interview User Study (55 participants) | Satisfaction84.41 | 4 |