Abstract :
The rapid adoption of multi-agent AI systems— ranging from prescriptive, workflow-driven deployments to fully agentic, autonomous ecosystems—raises urgent challenges for trust, accountability, and regulatory compliance. This paper introduces the TRACE Framework (Trust, Review, Accountability, Critique, Explainability), a governance-first architecture designed to make multi-agent AI systems auditable, policy-aligned, and operationally reliable across varying degrees of agent autonomy. TRACE embeds governance anchors at the agent level, enforces data privacy and policy checks, supplies a dedicated Critic agent for meta-validation, and preserves human-in- the-loop oversight where required. We present a layered architecture that separates Governance & Compliance, Operational Agents, and Oversight & Assurance, and provide a concrete methodology for instrumenting agent behaviour with provenance, explainability outputs, and per-agent metrics. A formal scoring rubric—comprising agent operational metrics, critic checks, and aggregation rules—yields an Overall System Confidence (OSC) that drives automated actions, human escalation, and continuous learning. Finally, we propose a suite of operational KPIs for each layer as Governance and Compliance Indicators (GCI), Agentic Performance Metrics (APM), and Assurance Indicators (AI) that enable financial institutions and other regulated organisations to deploy multi-agent systems that are efficient, auditable, and compliant. TRACE bridges the gap between regulatory expectations and system engineering practice— providing a practical roadmap for trustworthy multi-agent AI deployment in high-stakes domains.
Keywords :
Agentic AI, AI Governance, Explainable AI, Multi-agent systems, TRACE Framework, Trusted AIReferences :
1. European Commission. (2025). AI act. https://digitalstrategy.ec.europa.eu/en/policies/regulatory-framework-ai
2. International Organization for Standardization. (2023). ISO/IEC 42001:2023 – Artificial intelligence management system (AIMS) – Requirements (Tech. Rep. ISO/IEC 42001:2023). https://www.iso.org/standard/81230.html
3. National Institute of Standards and Technology. (2023). Artificial intelligence risk management framework (AI RMF 1.0) (NIST AI 100-1). https://nvlpubs.nist.gov/nistpubs/ai/nist.ai.100-1.pdf
4. Singla, A., Sukharevsky, A., Yee, L., Chui, M., & Hall, B. (2025). The state of AI: How organizations are rewiring to capture value. McKinsey & Company. https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai
5. Stanford Institute for Human-Centered Artificial Intelligence (HAI). (2025). 2025 AI index report. https://hai.stanford.edu/ai-index/2025-ai-index-report
6. Gartner. (2024). Tackling trust, risk and security in AI models (AI TRiSM). https://www.gartner.com/en/articles/ai-trust-and-ai-risk
7. Wang, T., et al. (2023). AutoGPT: Autonomous GPT-based agents for task automation. arXiv. https://arxiv.org/abs/2304.03442
8. Zhou, L., et al. (2023). AgentBench: Evaluating LLMs as agents. arXiv. https://arxiv.org/abs/2308.03688
9. Zheng, Y., et al. (2024). Emergent cooperation in LLM-based multi-agent systems. arXiv. https://arxiv.org/abs/2310.01985
10. Biran, O., & Cotton, C. (2022). Explainability and trust in autonomous systems. IEEE Transactions on Human-Machine Systems, 52(5), 801–813.
11. Biran, O., & Cotton, C. (2022). Explainability and trust in autonomous systems. IEEE Transactions on Human-Machine Systems, 52(5), 801–813.
12. Mehrabi, S., et al. (2022). A survey on bias and fairness in machine learning. ACM Computing Surveys, 54(6), 1–35.
13. Barbera, M. (2025). AI privacy risks & mitigations—large language models (LLMs). European Data Protection Board. https://www.edpb.europa.eu/system/files/2025-04/ai-privacy-risks-and-mitigations-in-llms.pdf
14. Cloud Security Alliance (CSA). (2025, February 6). MAESTRO: Agentic AI threat modeling framework. https://cloudsecurityalliance.org/blog/2025/02/06/agentic-ai-threat-modeling-framework-maestro
15. Feng, J., et al. (2025). AAGATE: Aligning agentic AI governance with NIST RMF principles. arXiv. https://arxiv.org/abs/2510.25863
16. Tian, Y., Luo, A., Du, J., Xian, X., Specht, R., Wang, G., Bi, X., Zhou, J., Srinivasa, J., Kundu, A., et al. (2025). An outlook on the opportunities and challenges of multi-agent AI systems. arXiv. https://arxiv.org/abs/2505.18397
17. Yang, Y., Peng, Q., Wang, J., & Zhang, W. (2024). Multi-LLM-agent systems: Techniques and business perspectives. arXiv. https://arxiv.org/abs/2411.14033
18. Acharya, B., Kuppan, K., & Divya, B. (2025). Agentic AI: Autonomous intelligence for complex goals–a comprehensive survey. IEEE Access.
19. Raza, S., Sapkota, R., Karkee, M., & Emmanouilidis, C. (2025). TRiSM for agentic AI: A review of trust, risk, and security management in LLM-based agentic multi-agent systems. arXiv. https://arxiv.org/abs/2506.04133
20. National Institute of Standards and Technology. (2024). Artificial intelligence risk management framework: Generative artificial intelligence profile. NIST Trustworthy and Responsible AI.
21. Chen, S., Liu, Y., Han, W., Zhang, W., & Liu, T. (2025). A survey on LLM-based multi-agent system: Recent advances and new frontiers in application. arXiv. https://arxiv.org/abs/2412.17481
22. European Union. (2016). General Data Protection Regulation (GDPR) – Article 25: Data protection by design and by default. https://gdpr-info.eu/art-25-gdpr/
23. Lee, M., & Tiwari, M. (2024). Prompt infection: LLM-to-LLM prompt injection within multi-agent systems. arXiv. https://arxiv.org/abs/2410.07283
24. Hannebauer, M. (1999). From formal workflow models to intelligent agents. In Proceedings of the AAAI-99 Workshop on Agent Based Systems in the Business Context (pp. 19–24).

